In this paper, we propose a criterion based on the variance variation of the sample eigenvalues to correctly estimate the number of significant components in high-dimensional principal component analysis (PCA), and it corresponds to the number of significant eigenvalues of the covariance matrix for p-dimensional variables. Using the random matrix theory, we derive that the consistent properties of the proposed criterion for the situations that the significant eigenvalues tend to infinity, as well as that the bounded significant population eigenvalues. Numerical simulation shows that the probability of estimator is correct by our variance variation criterion converges to 1 is faster than that by criterion of Passemier and Yao [Estimation of the number of spikes, possibly equal, in the high-dimensional case. J. Multivariate Anal., (2014)](PYC), AIC and BIC under the finite fourth moment condition as the dominant population eigenvalues tend to infinity. Moreover, in the case of the maximum eigenvalue bounded, once the gap condition is satisfied, the rate of convergence to 1 is faster than that of PYC and AIC, especially the effect is better than AIC when the sample size is small. It is worth noting that the variance variation criterion significantly improves the accuracy of model selection compared with PYC and AIC when the random variable is a heavy-tailed distribution or finite fourth moment not exists.