The short answer, from a frequentist perspective, is that the minimum number of flips that could possibly sway you against the coin being fair is obtained if all the flips are heads (or, equally, if all the flips are tails). The probability of this happening with a fair coin over $n$ flips is $2^{-n}$, and this first dips below $5$ percent at $n = 5$ flips. This is equivalent to a one-sided test for significance. If you require a two-sided test, then the probability of the flips being either all heads or all tails is $2^{-(n-1)}$, and this first drops below $5$ percent at $n = 6$ flips.
From a broader perspective, this question gets at the frequentist vs Bayesian debate. A Bayesian—to use the word attributively with a person—would say that the above analysis ignores any prior belief about the coin. If you look at the coin, and it has both a heads side and a tails side, and if you know you've flipped it $1000$ times before and it came down heads, say, $518$ times—consistent, under any reasonable assumption, with the coin being fair—you'll approach the event of it having flipped $5$ heads in a row quite differently than if you had never seen the coin before (including its other side).
To a Bayesian, the question is not addressed by seeing if we can falsify, at some level of statistical significance, the null hypothesis of the coin being fair. Instead, they would examine how the evidence of a sequence of flip outcomes adjusts any prior understanding about the coin's behavior.
They might, for example, start with a so-called uniform prior (meaning a presumption that the coin's probability of heads is uniformly distributed in the interval $[0, 1]$). With each flip, they update the distribution in accordance with Bayes's Theorem.
It turns out that these updates lead to the so-called Beta distributions. If you follow that link, you may notice that the Beta distribution—which includes the uniform distribution as a special case—is a conjugate prior for the Bernoulli distribution, which among other things is a good model for independent and identically distributed coin flips. This simply means that the math works out nicely: If you have a Beta distribution as the prior, then after a flip, the new a posteriori distribution is also Beta.
To be sure, there's no particular reason to favor the uniform prior, other than it plays nice. Other prior distributions might make more sense, depending on what you already know about the coin. That is in large part why the Bayesian perspective says the question can only be fully answered by stating the connection between what you believe of the coin before the experiment and what you believe of it afterward. If you are $100$ percent sure the coin was fair beforehand, then no length of experiment can every convince you otherwise (as lulu pointed out). You will only ever view it as progressively more strangely unlikely behavior of what you know with certainty to be a fair coin.
Incidentally, I would guess that you got $3.84$ because the count is too low for the normal approximation to be valid. The situation is small enough in scale that it's much more accurate to just rely on the binomial distribution.