Calculate the following gradient
$$\nabla_W \left( x^TW^{-T}W^{-1}x \right)$$
where $W$ is a $\mathbb{R}^{d×d}$ matrix and $x$ is a $\mathbb{R}^d$ vector. The result should be a $\mathbb{R}^{d×d}$ matrix.
I wonder whether there is a clean and compact form of the result. I first tried to write this as
$$2(W^{-1}x)\frac{d(W^{-1}x)}{dW}$$
but the latter one is a tensor (actually it should be $2 \sum_{i=1}^d(W^{-1}x)_i\frac{d(W^{-1}x)_i}{dW}$ but the next step is messy.