Problem:
Given $\alpha\in\mathbb{R}^+$, $N\in \mathbb{Z}^+$ and $D\in \{N, N+1, N+2, \cdots\}$, a random matrix $\mathbf{A}$ is generated by the following steps:
$(1)$ Randomly select $N$ numbers from $\{1,2,\cdots,D\}$ to form a sequence $p=\{p_i\}_{i=1}^N$.
$(2)$ Then calculate $\mathbf{A}=[a_{ij}]_{N\times N}$, where $a_{ij}=e^{-\alpha |p_i - p_j|}$.
Please prove or disprove the following proposition:
$\mathbf{A}$ converges to $\mathbf{I}_N$ in probability, i.e., for any $\epsilon>0$ and choice of norm $\|\cdot\|$, there is: $$ \mathbb{P}[\|\mathbf{A}-\mathbf{I}_N\|\geq\epsilon]\to0~~(D\rightarrow \infty). $$
My Efforts:
I am confused by how to start.
I may know that the diagonal elements of $\mathbf{A}$ will be all ones, since $|p_i-p_i|=0$.
And I may know that all elements of $\mathbf{A}$ are in $[0,1]$ and $\mathbf{A}$ is symmetric.
Intuitively, I guess that when $D$ increases, the absolute distances between each two $p_i$s may become larger and larger, so $a_{ij}$ is expected to be smaller and smaller.
I also write the following Python program for numerical validation:
import numpy as np
import random
from scipy import spatial
alpha = 1
N = 10
I = np.eye(N)
for D in range(N, 10000):
MSE = 0.0
for i in range(100):
p = np.array(random.sample(range(1, D + 1), N)).reshape(N, 1)
A = np.exp(-alpha * spatial.distance.cdist(p, p))
MSE += np.sum((A - I) ** 2.0)
MSE /= (100 * N * N)
print(MSE)
I can see that when $D$ increases, the mean squared error between $\mathbf{A}$ and $\mathbf{I}_N$ converges to zero.
0.027683220252563596
0.02508590350202309
0.02317795057344325
...
0.0001934704436327538
0.00032059290537374806
0.0003270223508894337
...
5.786435956425624e-05
1.1065792791574203e-05
5.786469182583059e-05
How to prove/disprove the proposition by exactly analysing the process of $D\rightarrow \infty$?