The cross naturally subdivides into five parts, namely left, right, top, bottom, middle. That's not convenient. Let's consider four parts, that is we take the entire vertical bar, which we split in two, and add the left part and the right part.
Each of these parts is rectangular, so you can obtain them from $u$ and $v$ as you described, by stretching to the desired length and height individually.
However, we have to pay attention that the total masses match.
So, for the cross, compute the entire area, and then the four individual areas.
This gives you four ratios $p_1,\dots,p_4\in[0,1]$ that add up to $1$, a probability mass function. Notice that $p_1=p_2$ and $p_3=p_4$, since our splits were symmetric.
Thus, we can for example split $u\in[0,0.5)\cup[0,5,1)$ and $v\in[0,2p_1)\cup[2p_1,1)$. This gives you a split of the unit square into four parts, all with the correct individual total mass, matching their counterparts in the cross. The last step is to map the individual rectangles in the unit square to the corresponding target rectangles, by stretching them as you described. How do we implement the above? Let $C=R_1\cup R_2\cup R_3\cup R_4$ be the cross composed of the $R_1$, the upper half of the vertical bar, $R_2$, the lower half of the vertical bar, $R_3$ the left part, and the right part $R_4$. Let $x_1<x_2<x_3<x_4$ be the relevant $x$-coordinates, $y_1<y_3<y_4<y_5$ the relevant $y$-coordinates, and let $y_2=0.5y_1+0.5y_5$ be the split of the vertical bar.
Then we have $R_1=[x_2,x_3]\times[y_2,y_5]$, $R_2=[x_2,x_3]\times[y_1,y_2)$, $R_3=[x_1,x2)\times[y_3,y_4]$ and $R_4=(x_3,x_4]\times[y_3,y_4]$. Let $w_1=x_3-x_2$ be the width of $R_{1,2}$, and $h_1=y_5-y_2=y_2-y_1$ their height. Let $w_2=x_2-x_1=x_4-x_3$ and $h_2=y_4-y_3$ be the lengths for $R_{3,4}$. Thus, the total area is $A=2w_1h_1+2w_2h_2$, and the ratios are $p_1=p_2=w_1h_1/A$, $p_3=p_4=w_2h_2/A$.
Now, we set
\begin{align*}
(x,y)=&\unicode{120793}\{u<0.5,v<2p_1\}(x_2+2w_1u,y_2+\frac{h_1}{2p_1}v)+\\
&\unicode{120793}\{u\ge 0.5,v<2p_1\}(x_2+2w_1u,y_1+\frac{h_1}{2p_1}v)+\\
&\unicode{120793}\{u<0.5,v\ge 2p_1\}(x_1+2w_2u,y_3+\frac{h_2}{1-2p_1}v)+\\
&\unicode{120793}\{u>0.5,v\ge 2p_1\}(x_3+2w_2u,y_3+\frac{h_2}{1-2p_1}v).
\end{align*}
The probability to end up in $R_1$ and $R_2$ is $0.5\cdot 2p_1=p_1$ each.
The probability to end up in $R_3$ and $R_4$ is $0.5\cdot (1-2p_1)=0.5\cdot 2p_3=p_3$ each, so that's good.
Now, conditional to being in $R_1$ (meaning given the event $u<0.5, v<2p_1$), $u$ and $v$ are still independent and uniform. But then also the resulting coordinates are independent and uniform, because the first one only depends on $u$ and the second one only on $v$, and both are affine transformations. For the first coordinate we notice that it reaches from $x_2$ to $x_2+2w_1\cdot 0.5=x_3$, as desired. The rest follows analogously.
This shows that conditional to being in $R_i$ we're uniform on $R_i$, and the probability of ending up in $R_i$ is correct. One last question remains: If $z\in C$ is uniform, is it true that $z\in R_i$ is uniform, given that $z\in R_i$? Yes, it is, and we obtain that by integration, meaning
$$\mathbb P(z\in\mathcal E|z\in R_1)=\frac{\mathbb P(z\in\mathcal E\cap R_1)}{\mathbb P(z\in R_1)}=\frac{\frac{1}{A}\int_{\mathcal E\cap R_1}\mathrm d z}{\frac{1}{A}w_1h_1}
=\frac{1}{w_1h_1}\int_{\mathcal E\cap R_1}\mathrm d z.$$
The idea, frequently used in computer science, is that you can obtain a random variable by inverting the cumulative distribution function.
In our case, let $A=\int_S\mathrm dx$ be the area of $S$, and let $F_1:\mathbb R\rightarrow[0,1]$, $x\mapsto\frac{1}{A}\int_{S\cap((-\infty,x]\times\mathbb R)}\mathrm d x'$, be the cumulative distribution function of the marginal on the first coordinate (which is not uniform). For given $x\in\mathbb R$ let $S_x=\{y\in\mathbb R:(x,y)\in S\}$ be the section of $S$ at $x$, further let $A_x=\int_{S_x}\mathrm dy$ be the length of $S_x$, and let $F_{2,x}:\mathbb R\rightarrow[0,1]$, $y\mapsto\frac{1}{A_x}\int_{S_x\cap(-\infty,y]}\mathrm{d}y'$ for all $x\in\{x'\in\mathbb R:A_{x'}>0\}$. The second distribution is the conditional probability given that the first coordinate is $x$, in the kernel sense.
Notice that $F_1$ is invertible on $[x_-,x_+]$, where $x_-=\max\{x:F_1(x)=0\}$ and $x_+=\min\{x:F_1(x)=1\}$. Let $X=F_1^{-1}(u)$, using $u,v$ i.i.d. uniform on $[0,1)$. Similarly, let $Y=F_{2,X}^{-1}(v)$, then $(X,Y)$ is uniform on $S$. Clearly, this is a fairly involved construction for this problem (and we haven't even yet shown that $(X,Y)$ is measurable), but when the target distribution is something that is not as accessible as the uniform distribution on $S$, then this method might prove helpful.