I'm currently trying to learn Monte Carlo Markov chain, and I'm having troubles to grasp what it is good for and hoping that someone could help me. I want to get some help of general understanding of MCMC, therefore the question are a bit general.
From my text book it is explained:
Given a probability distribution $\pi$, the goal of MCMC is to simulate a random variable $X$ whose distribution is $\pi$. Often, one wants to estimate an expectation or other function of a join distribution from a high dimensional space. The MCMC algorithm construct an ergodic Markov chain whose limiting distribution is the desired $\pi$. One then runs the chain long enough for it to converge, and outputs the final element or elements of the Markov sequence as a sample from $\pi$. MCMC relies on the face that the limiting properties of ergodic Markov Chains have some similarities to independent and identically distributed sequences.
From the lecture is it shortly summarized as:
-Start with a particular prob. distribution
-Define a Markov chain with the prob distribution as limiting distribution
-After many step, a value in simulation is distributed according to the limiting distribution.
-Idea is that it can be difficult to obtain sample otherwise.
My question:
Why use a Markov chain to sample from a known distribution instead of directly sampling from it? Isn’t that unnecessarily.
If direct sampling is difficult, wouldn’t constructing a Markov chain with the desired limiting distribution be even harder? How is this done, especially for high-dimensional distributions in MCMC?
Would really be grateful if someone could help me understand this, since I've both googled and searched YouTube for "simple enough" answers, before diving into the theory about this.
Best regards,