In this paper, we propose a novel neighbor-preferential growth (NPG) network model. Theoretical analysis and numerical simulations indicate the new model can reproduce not only a scale-free degree distribution and its...In this paper, we propose a novel neighbor-preferential growth (NPG) network model. Theoretical analysis and numerical simulations indicate the new model can reproduce not only a scale-free degree distribution and its power exponent is related to the edge-adding number m, but also a small-world effect which has large clustering coefficient and small average path length. Interestingly, the clustering coefficient of the model is close to that of globally coupled network, and the average path length is close to that of star coupled network. Meanwhile, the synchronizability of the NPG model is much stronger than that of BA scale-free network, even stronger than that of synchronization-optimal growth network.展开更多
The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effectiv...The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effective algorithm to estimate the finite mixture model parameters. However, EM algorithm can not guarantee to find the global optimal solution, and often easy to fall into local optimal solution, so it is sensitive to the determination of initial value to iteration. Traditional EM algorithm select the initial value at random, we propose an improved method of selection of initial value. First, we use the k-nearest-neighbor method to delete outliers. Second, use the k-means to initialize the EM algorithm. Compare this method with the original random initial value method, numerical experiments show that the parameter estimation effect of the initialization of the EM algorithm is significantly better than the effect of the original EM algorithm.展开更多
Consider the regression model Y=Xβ+ g(T) + e. Here g is an unknown smoothing function on [0, 1], β is a l-dimensional parameter to be estimated, and e is an unobserved error. When data are randomly censored, the est...Consider the regression model Y=Xβ+ g(T) + e. Here g is an unknown smoothing function on [0, 1], β is a l-dimensional parameter to be estimated, and e is an unobserved error. When data are randomly censored, the estimators βn* and gn*forβ and g are obtained by using class K and the least square methods. It is shown that βn* is asymptotically normal and gn* achieves the convergent rate O(n-1/3).展开更多
文摘In this paper, we propose a novel neighbor-preferential growth (NPG) network model. Theoretical analysis and numerical simulations indicate the new model can reproduce not only a scale-free degree distribution and its power exponent is related to the edge-adding number m, but also a small-world effect which has large clustering coefficient and small average path length. Interestingly, the clustering coefficient of the model is close to that of globally coupled network, and the average path length is close to that of star coupled network. Meanwhile, the synchronizability of the NPG model is much stronger than that of BA scale-free network, even stronger than that of synchronization-optimal growth network.
文摘The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effective algorithm to estimate the finite mixture model parameters. However, EM algorithm can not guarantee to find the global optimal solution, and often easy to fall into local optimal solution, so it is sensitive to the determination of initial value to iteration. Traditional EM algorithm select the initial value at random, we propose an improved method of selection of initial value. First, we use the k-nearest-neighbor method to delete outliers. Second, use the k-means to initialize the EM algorithm. Compare this method with the original random initial value method, numerical experiments show that the parameter estimation effect of the initialization of the EM algorithm is significantly better than the effect of the original EM algorithm.
文摘Consider the regression model Y=Xβ+ g(T) + e. Here g is an unknown smoothing function on [0, 1], β is a l-dimensional parameter to be estimated, and e is an unobserved error. When data are randomly censored, the estimators βn* and gn*forβ and g are obtained by using class K and the least square methods. It is shown that βn* is asymptotically normal and gn* achieves the convergent rate O(n-1/3).