I am trying to understand how components of a vector in the dual basis transform covariantly as mentioned in this quote.
If you seek to define a quantity (such as vector A) that remains invariant under a transformation of coordinates, you have a choice: you can combine superscripted (contravariant) components with subscripted (covariant) basis vectors, or you can combine subscripted (covariant) components with superscripted (contravariant) basis vectors.
Moreover I am trying to understand what transforms covariantly and contravariantly means exactly.
In the case of superscripted (contravariant) components with subscripted (covariant) basis vectors what I have inferred from the answers to my previous question is the following. Consider $V=\mathbb{R}^n$ and let $B$ denote the matrix of old basis vectors and $B'$ the matrix of new basis vectors. Then
$$B' = \begin{pmatrix}\vert & \vert & & \vert \\ b'_1 & b'_2 & \ldots & b'_n\\ \vert & \vert & & \vert \end{pmatrix} = BP = \begin{pmatrix}\vert & \vert & & \vert \\ b_1 & b_2 & \ldots & b_n\\ \vert & \vert & & \vert \end{pmatrix} P $$
and $[v]_{B'} = P^{-1}[v]_{B}$. Then contravariant means the use of the inverse of the matrix used for covariant. Hence the appearance of $P^{-1}$ and $P$. But this terminology of contravariant and covariant doesn't indicate whether the matrix $P^{-1}$ or $P$ multiples on the left or the right.
Now consider the the case of viewing the vector as (covariant) components with superscripted (contravariant) basis vectors. The matrix of dual basis vectors corresponding to $B$ is $D=(B^{-1})^T$. The matrix of dual basis vectors of $B'$ is $D' = ((B')^{-1})^T= ((BP)^{-1})^T=(B^{-1})^T(P^{-1})^T$. So that $D' = D(P^{-1})^T$ and $[v]_{D'} = P^T[v]_{D}$. Although the $P$ and $P^{-1}$ have switched places compared to the above as expected, I'm confused as there are transposes and yet the covariant contravariant terminology says nothing about this. Unless maybe I'm suppose to view the (covariant) components with superscripted (contravariant) basis vectors case as transposed so that all equations are transposed with everything written as row vectors so as to eliminate the transpose on $P$.
Surprisingly none of the books I've looked at talk about whether or when the matrix $P^{-1}$ or $P$ multiples on the left or the right or when to view things maybe as row vectors. Maybe I'm just looking at the wrong books?