Let $I$ denote the unit interval and $\mu$ be the Lebesgue measure. Let $S:I\to I$ be the map defined as $S(x)=2x \pmod{1}$. Then it is known that for any measurable subset $A$ of $I$ we have $$ \lim_{n\to \infty} \mu(S^{-n}A\cap B) = \mu(A)\mu(B) $$
So if we fix a subset $A$, we see that $A$ gets "mixed out" in the interval by backwards iterates of $S$.
I would like to say that $A$ had a certain entropy in the beginning, but as $n$ increases the entropy of $S^{-n}(A)$ increases.
It's like what we read in not-so-rigorous thermodynamics: Start with a box with a partition. On one side of the partition is a gas, and the other side is vacuum. Once the partition is removed the gas takes up all the space. The entropy of the gas (whatever that means) increases in the process.
So my question is: Is there a rigorous formulation of the the notion of entropy I was trying to hint at in my example? Also, can the notions of ergodicity and mixing be recast in the language of entropy?
Thank you.