I'll try to explain what the above graphs represent. The histogram in
the first one is that of the eigenvalues of a sample covariance matrix
(s.c.m.) formed from 4000 samples of a 200 dimensional random vector whose
population covariance matrix has 3 distinct eigenvalues:
1, 3, and 10, with respective multiplicities 40, 80, and 80. The
assumptions on the random vector are general enough to include multivariate
normals. The function you see in both graphs is the limiting density
(whose existence and properties are part of my work) of the empirical
distribution of the eigenvalues of s.c.m. as the dimension
and sample size approach infinity, the population eigenvalues remain at
(1,3,10) with relative proportions (.2,.4,.4), and the ratio of dimension
to sample size is .05. The closeness of the limiting density to the
histogram displays several things: 1) the validity of the above mentioned
``limit theorem'', 2) the ``curse of high dimensionality'', namely that, even
though 4000 samples were used to estimate the population matrix, because
of the dimension of the random vector, the estimation is not so great, and
3) since the histogram of eigenvalues is predictable from the limit
theorem and because of an understanding of how the limiting
density should look, due to the relatively small ratio of dimension
to sample size (.05), the histogram does reveal quite a bit of information
on the eigenvalues of the population matrix (after all, there are 3 clumps
centering around the population eigenvalues).
The second graph also contains a scatter plot of all 200 eigenvalues,
in clumps of (from left to right) 40, 80, and 80, demonstrating
Z.D. Bai and my previous results (see references 32 and 33 in publications): no
eigenvalues where they shouldn't be, and the
right number where they should be. These two facts,
although thoroughly reasonable from the above limit theorem, are a bit
more complicated to prove. The limit theorem only guarantees proportion
of eigenvalues are converging to the right numbers.