I'm currently learning machine learning, and I came across this equation called Least Squares Regression.
X and w are both matrices. The multiplication of both matrices becomes y hat, which is theoretically supposed to be equal to y.
We want to minimize the squared error given by this equation by changing w.

w can be solved by a derivation of the function to w, and setting the equation to zero.
The question is, what does it intuitively mean?
I know that in derivative, we are trying to find the rate of change. BUT what does it mean the rate of change = 0 intuitively?