Abstract
In this paper, we consider the problem of fitting a sparse precision matrix to multivariate Gaussian data. The zero elements in the precision matrix correspond to conditional independencies between variables. We focus on the estimation of a class of sparse precision matrix which represents the scale-free networks. It has been demonstrated that some of the important networks display features similar to scale-free graphs. We propose a new log-likelihood formulation, which promotes the sparseness of the precision matrix as well as the topological structure of scale-free networks. To optimize this new energy formulation, the alternating direction method of multipliers form is used with the general \(L_1\)-regularized loss optimization. We tested our proposed method on various databases. Our proposed method exhibits better estimation performance with various number of samples, N, and different selection of sparsity parameter, \(\rho \).
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Undirected graphical models study the relationship between a set of variables. The Gaussian networks are a way of study to extract undirected relationships among the data. In the Gaussian networks, the variables are assumed to follow the multivariate Gaussian distribution with mean, \(\mu \), and covariance matrix, \(\varSigma \).
Gaussian models aim to find out the dependencies between variables. The relationship between variables in these networks can be represented using the precision matrix (\(\varOmega \)) of data sets and a graph in which the nodes and edges represent the variables and their conditional dependencies, respectively. The estimation of precision matrices from data is a problem of many applications and fields such as health, finance and biology. Zero elements reveal the conditional independencies of corresponding variables. If the ijth component of \(\varOmega \) is zero, then variables i and j are conditionally independent, given the other variables.
The problem formulation can be summarized as follows: Let D represent a data set of cases, Z represent a set of variables, \(B_s = (Z,E)\) be a graph representing interactions between components of Z. \(B_{s}\) is any set in all possible networks, B. If E consists of an edge between these two variables, \(Z_i\) and \(Z_j\), then these variables are said to be dependent. The objective is to find the best network which maximizes \(P(B_{s}|D)\) or \(P(B_{s},D)\). In another definition, the aim is to estimate the best network \(B_{s}\) which can be described as follows:
P(B|D) is modeled using the Gaussian distribution in the Gaussian networks. The inverse of the covariance matrix can be estimated by the log-likelihood estimation with respect to the covariance matrix as follows:
where \(\varOmega =\varSigma ^{-1}\) is the precision matrix and \({\mathbf{tr}}\) denotes the trace.
The number of nonzero elements in the precision matrix and number of edges in the structure are equivalent measures of complexity. Several methods have been proposed to obtain the sparse precision matrix and tackle the existing challenges. For instance, Dempster [2] proposed the idea of the precision matrix estimation. The number of edges in the structure, the number of parameters and the number of nonzero elements in the precision matrix are equivalent measures of complexity in the Gaussian graphical models [3]. Therefore, many studies have been focusing on the sparseness of the precision matrix recently. In the literature, there are various derivations of the following likelihood formulation to estimate the sparse precision matrix:
where \(\varOmega =\varSigma ^{-1}\) is the precision matrix, \({\mathbf{tr}}\) denotes the trace, \(\rho >0\) is a scalar parameter which controls the size of the penalty, hence the sparsity of the solution, \(\Vert \varOmega \Vert \) is the \(L_1\) norm which is the sum of absolute values of \(\varOmega \).
Meinshausen and Buhlmann [4] proposed a method to perform edge selection at each node in the graph using the least absolute shrinkage and selection operator (LASSO) which was proposed by Tibshirani [5]. They demonstrated that their estimate is shown to be both more accurate and computationally much more efficient than the traditional forward selection maximum-likelihood estimation (MLE) strategy. The forward MLE estimation has poor accuracy when the number of nodes in the graph is comparable to the number of observations. Furthermore, when the number of observations is remarkably less than the number of nodes, the empirical covariance \(\varSigma \) is singular so that we cannot access the required information about interactions between nodes.
In this context, several algorithms have been proposed such as Glasso (by Friedman et al. [6]), SPICE (by Rothman et al. [7]), SCAD (by Fan et al. [8]), COVSEL (by Banerjee et al. [9]) and CLIME (by Cai et al. [10]). Friedman et al. [6] estimated sparse graphs by a lasso penalty applied to the precision matrix. They proposed a well-known algorithm, named as Glasso. They used the blockwise coordinate descent approach in Banerjee et al. [9] as a launching point and proposed a new algorithm for the rest of the problem solution. They started with a given covariance matrix and regenerated data from the given covariance. The new covariance matrix is estimated from the generated data. Then for each variable they solved the lasso problem (\(\min _{\beta }\{0.5\Vert \varSigma ^{1/2}_{11}\beta -b\Vert +\rho \Vert \beta \Vert _1\}\) where \(b=\varSigma ^{-1/2}_{11}s_{12}\), \(\varSigma =\left[ \begin{array}{ll} \varSigma _{11}~~\sigma _{12}\\ \sigma _{12}^T~~\sigma _{22} \end{array}\right] ,~S=\left[ \begin{array}{ll} S_{11}~~s_{12}\\ s_{12}^T~~s_{22} \end{array}\right] \)) using the blockwise coordinate descent method. This process produces a \(p-1\) vector solution \(\widehat{\beta }\) which is used to calculate \(\sigma _{12}=\sigma _{11}\widehat{\beta }\). This process continues until convergence. They stated that their algorithm is 30 to 4000 times faster than other competing methods.
Later, Dahl et al. [11] solved the likelihood estimation problem using Newton’s method and conjugated gradient method. Their proposed method solves problems where a sparse structure of \(\varSigma ^{-1}\) is known a priori. Recently, Honorio et al. [3] proposed a method promoting the variable selection in addition to the sparseness of the precision matrix. In this method the sparseness is imposed not only at the edge level but also at the important variable selection level. They applied the block coordinate descent method and compared the results with the graphical lasso and covariance selection methods. Yuan et al. [12] proposed a penalized-likelihood method which has the model selection and parameter estimation simultaneously in the precision matrix estimation. They ensured that the estimator of the precision matrix is positive definite. D’aspremont et al. [13] proposed a likelihood formulation which penalizes the cardinality of the precision matrix. To solve the cardinality penalty term, they derived a convex relaxation solution. Other techniques have been developed in the similar direction and can be found for instance in [14–16] and the references therein.
One of the major limitations in traditional precision matrix estimation methods is that they aim the sparsity uniformly on each variable. In reality, however, most networks display scale-free properties [17]. Hence, the traditional methods would give poor performance to estimate the topology of the specific networks. Hub-dominated networks are usually dominated by a relatively small number of nodes (hubs) which are connected to many other nodes, and these networks are considered resistant to accidental failures but extremely vulnerable to coordinated attacks. Some examples of such data structures can be described as the interactions between research collaborations, accesses to the World Wide Web, role plays in Hollywood, etc. While many studies have been carried out for sparse network learning, little is done on the learning of the sparse Gaussian graphical models aiming to preserve properties of networks which are believed to be scale-free or have dominating hubs. We prefer to name both topologies as ‘scale-free’ networks for simplicity in this paper.
Liu et al. [18] proposed an \(L_1\) regularization formulations, where coefficients of nodes with higher degree are decreased. They proposed a formulation to promote the features of hub-dominated networks. One of the drawbacks of this paper is that the proposed formulation is nonconvex which does not guarantee the convergency. Also, they tested their method on relatively very small data sets (e.g., \(p<120\)). In this paper our problem can be described as follows: Given a set of data, we solve a convex likelihood formulation to make the resulting graph as sparse as possible, whereas the hub-dominated graphical topology is protected.
Recently, analyzing and using scale-free topology and graphical models improves the computational costs and accuracy. For instance, Liu [19] employed the scale-free network to represent the inter-individual interactions in particle swarm optimization (PSO) algorithm. They reported that in contrast to the traditional PSO with fully connected topology or regular topology, the scale-free topology incorporates the diversity of individuals in searching and information dissemination ability, leading to a quite different optimization process. They obtained results with respect to several standard test functions. The outcome demonstrated that their updated PSO algorithm gives better balance between the convergence speed and the optimum quality than the traditional PSO algorithms. They further explored that the cooperation of hub and regular nodes plays a crucial role in optimizing the convergence process. Zhang and Huang [20] studied the combined impact of reinstalling system and network topology on the spread of computer viruses over the Internet. Their experiments show that the virus-free equilibrium is globally asymptotically stable when its spreading threshold is less than one; nevertheless, it is proved that the viral equilibrium is permanent if the spreading threshold is greater than one. The impacts of different model parameters on spreading threshold are analyzed. Haoran [21] proposed a new scale-free topology model which has both fault tolerance against random faults and intrusion tolerance against selective remove attacks at the same time. The optimal scale-free topology which can keep the fault tolerance and maximize intrusion tolerance is obtained through analyzing the effect of topological degree distribution on these properties of topological fault tolerance and topological intrusion tolerance. The simulation results show that their method also can reduce their fragility for selective remove attacks and further prolong its lifetime.
The major contributions of this paper can be listed as follows: (i) We propose a novel structure learning method which is capable of working on the scale-free networks. (ii) The proposed convex formulation emphasizes the sparsity of the precision matrix and features of the scale-free topology. (iii) The proposed method is able to guard the hubs since hub-dominated networks are considered extremely vulnerable to coordinated attacks. One may diminish the connections on the hubs when the penalty constant is chosen for only the sparsity of the precision matrix.
The rest of the paper is organized as follows. In Sect. 2, the basic formulation of the precision matrix estimation problem and the proposed method is presented. We show the numerical performance of the proposed method and some of alternatives in Sect. 3 using synthetic and real data sets. We conclude our paper in Sect. 4.
2 Method
In this paper, we use the alternating direction method of multipliers (ADMM) form [22, 23] to solve our proposed log-likelihood formulation [1]. First, we start to explain and derive a simple \(L_1\) problem using ADMM solution. For a generic \(L_1\) problem, the objective is written as follows:
where l is any convex loss function. In ADMM form, this problem is written as
where \(g(y)=\rho \Vert y\Vert _1\). The new updates are estimated as follows:
where the soft thresholding operator, S, is defined as follows:
In a scale-free network, some nodes, which are named as hubs in this paper, have a tremendous number of connections to other nodes, whereas most nodes, which are named as normal or regular nodes, have just a few. The hubs can have considerably much larger number of connections with respect to the normal nodes (e.g., hundreds, thousands or even millions of links). Therefore, these networks appear to have no scale [17].
In our problem the variables can be defined as \(Z=\{Z_r\cup Z_h\}\) where \(Z_r\) and \(Z_h\) represent the sets of the normal variables and hubs, respectively. The number of variables in \(Z_r\) and \(Z_h\) classes can be represented as \(p_r\) and \(p_h\), respectively. We assume that the precision matrix consists of two classes as \(\varOmega =[\varOmega _r+\varOmega _h]\) where \(\varOmega \in R^{px p}\), \(\varOmega _r\in R^{p\text {x}p}\) and \(\varOmega _h\in R^{p\text {x}p}\).
We propose a new log-likelihood formulation which promotes the sparseness of the precision matrix and features of the scale-free topology. Given a sample covariance matrix \(\varSigma \), we estimate a precision matrix \(\varOmega ^*\) for p variables as follows:
where \(Y_r\in R^{p\text {x}p}\), \(Y_h\in R^{p\text {x}p}\), \(G\in R^{p\text {x}p}\), \(G=[\rho _{ij}]\), \(i=\{1,2,\ldots ,p\}\), \(j=\{1,2,\ldots ,p\}\) and \(\rho _{ij}=\frac{\rho _i+\rho _j}{2}\). The augmented Lagrangian (using the scaled dual variable) can be written as:
where \(\kappa =1\) in our method. The scaled form of ADMM for this problem is:
where k is the iteration number. Using the scaled form of ADMM, the optimization steps of \(\varOmega \), Y and U can be described in the following three subsections.
Synthetic scale-free networks with a \(p=16\), b \(p=30\). c One of the real scale-free network [24] with \(p=204\)
a Sensitivity of 30 connections between 6 hubs and other variables in the scale-free network with \(p=16\) and \(N=1600\). b Histogram of the variable connections on the data set with \(p=16\). c Sensitivity of 19 connections between 3 hubs and other variables in the scale-free network with \(p=30\) and \(\rho =.045\). d Histogram of the variable connections on the data set with \(p=30\). e Sensitivity of the hubs in the scale-free network with \(p=204\) and \(N=20{,}400\). f Histogram of the variable connections on the data set with \(p=204\)
\(\varOmega \mathbf {-optimization:}\)
where \(\varOmega =[\varOmega _r+\varOmega _h]\), \(Y=[Y_r+Y_h]\) and \(U=[U_r+U_h]\). After this point, \(\varOmega \)-optimization step can be solved using an analytic method. The first-order optimality condition is that the gradient should vanish as follows:
\(\varOmega \) is constructed to satisfy this condition and thus minimizes the \(\varOmega \)-minimization objective. First, take the orthogonal eigenvalue decomposition of the right-hand side,
where \(\varLambda \) is the diagonal matrix which contains the eigenvalues of \(\kappa (Y^k-U^k)-\varSigma \), and Q is the matrix which contains eigenvectors of \(\kappa (Y^k-U^k)-\varSigma \). Note that \(QQ^T=Q^TQ=I\) where I is the identity matrix. Let us multiply the right side of Eq. 14 with \(Q^T\) on the left and Q on the right side as follows:
where \(\widetilde{\varOmega }=Q^T\varOmega Q\). We can now construct a diagonal solution of this equation, i.e., find positive numbers \(\widetilde{\varOmega }_ii\) that satisfy \(\kappa \widetilde{\varOmega }-\frac{1}{\widetilde{\varOmega }}=r_{i}\), where \(r_{i}\) is the corresponding eigenvalue. By the quadratic formula, each elements can be estimated as \(\widetilde{\varOmega }_{ii}=\frac{r_i+\sqrt{r_i^2+4\kappa }}{2\kappa }\), where the solution is always positive since \(\kappa >0\). It follows that \(\varOmega =Q\widetilde{\varOmega }Q^T\) satisfies the optimality condition \(\kappa (Y^k-U^k)-\varSigma =Q\varLambda Q^T\), so this is the solution to the \(\varOmega \)-minimization.
Y-optimization
One can easily find that
where \(\rho _r\) and \(\rho _h\) are the specific constants assigned to \(Z_r\) and \(Z_r\), respectively. In this problem, we need to ensure that the connections on the hubs will not be lost in the scale-free networks when we seek to have a sparse precision matrix. Therefore, we suggest to use two penalty parameters for the scale-free networks. We offer that constants for the hubs can be accepted as \(\rho _h=-\rho _r/c\) where we choose \(c=N\) in our method. This representation of the proposed formulation aims to minimize \(||\varOmega _r||_1\), which promotes the sparseness, and to maximize \(||\varOmega _h||_1\), which promotes the presence of hubs. Hence, the scaled form of ADMM for this problem can be written as follows:
Since
the final Y-optimization can be written as follows:
which corresponds to \(Y^{k+1}=[Y_r^{k+1}+Y_h^{k+1}]\) and where S is the soft thresholding operator.
U-optimization The scaled form of ADMM for this problem is:
We can accept this formulation as follows:
which corresponds to \(U^{k+1}=[U^{k+1}_r+U^{k+1}_h]\). Our method is summarized as in Algorithm 1.

3 Numerical results
To evaluate the proposed method, we use structures of given scale-free networks that have the ground truth of the precision matrix. The precision matrix is normalized using the similar approach described in [8]. Nonzero off-diagonal elements in \(\varOmega \) are regenerated uniformly from intervals \([-1:-0.5]\bigcup [0.5:1]\). The value of each diagonal element is set as a factor of the sum of the absolute values of its corresponding row elements. A constant value is selected to ensure the positive definiteness of \(\varOmega \). Finally, each row is divided by its corresponding diagonal entry so that the precision matrix has diagonal values of ones. Using this precision matrix, the covariance matrix, \(\varSigma \), is obtained to generate various number of samples. We simulate data sets each with various number of samples i.i.d. generated from multivariate Gaussian distribution \(N(0, \varSigma ^{-1})\).
The results of the proposed and some of the state-of-the-art methods are evaluated with different formulations such as sensitivity, specificity and Matthew correlation coefficient (MCC) defined as follows:
where TP, TN, FP, FN are the number of true positives, true negatives, false positives, false negatives, respectively. Note that \(TP_{HUBS}\) and \(FN_{HUBS}\) represent true positives and false negatives on hubs’ connections, respectively. The bigger sensitivity, specificity and MCC indicate better estimation.
The proposed method is compared with Covsel [16] and Glasso [6] on both synthetic and real data sets. We test the proposed and alternative methods using various penalty constants and sample size. We choose some of the comparisons in this section. Note that the results are selected when Covsel [16] and Glasso [6] methods give their highest accuracy. The accuracy and robustness of the methods are shown using (i) the accuracy table, (ii) histograms analyzing connections between all nodes, (iii) sensitivity measurements on the connections between the hubs and other nodes, (iv) partial regions of the resulting networks of the real data networks.
We tested the methods on two synthetic and three real networks. Two synthetic networks are shown in Fig. 1a, b, whereas one of the real network is shown in Fig. 1c. The number of variables in these networks is \(p=16\), \(p=30\) and \(p=204\), respectively. The second and third real networks have “2361” and “4039” variables, respectively.
3.1 Synthetic data
First, we begin with two small synthetic examples to assess the method to recover the sparse and scale-free structures. In this experiment, we measure the robustness of the algorithms based on different penalty constants and sample size. Note that \(\rho =\rho _R\) in the experiments.
Figure 2a shows the sensitivity only on the connections between the hubs and other nodes under various penalty constants. Here, our method ensures (as the alternatives do not) that the connections on the hubs will not be lost as much as possible in the scale-free networks when we seek to have a sparse precision matrix under various penalty terms. Also, the results show that the proposed method is more robust to the changes. The number of hubs can be selected by users or the power-law distribution. The histogram of “the number of connections” with respect to “the number of variables” is shown in Fig. 2b. Here, we use an illustration to show how the proposed and other two methods estimate the connections between all variables, especially on the hubs (when the number of connections increases). Our proposed method is able to protect the hubs, whereas the sparseness of the precision matrix is warranted. Table 1(i) shows the results on this data set with a specific scenario.
A small region of the scale-free network [24] (as shown in Fig. 1c) with \(p=204\). This area shows only the 4 hubs and their connections with other nodes. a Ground truth of the variable connections. The results of the b Covsel, c Glasso and d our proposed algorithm. The red dashed color represents the FNs. These results are obtained when the sensitivity measurements of all 204 variables’ connections are 91, 90 and 97 % for Covsel [16], Glasso [6] and the proposed method, respectively (color figure online)
A similar experiment is applied on the network shown in Fig. 1b. Figure 2c shows the sensitivity only on the connections between the hubs and other nodes under various sample sizes. The histogram that shows “the number of connections” with respect to “the number of variables” is shown in Fig. 2d. In this figure, we can see that the numbers of hubs’ connections of our method and the ground truth are almost similar to each other. One of the most important contributions of the proposed method is to guard the hubs since scale-free networks are considered extremely vulnerable to coordinated attacks to hubs. Table 1(ii) shows the results on this data set with a specific scenario.
3.2 Real data
A considerably bigger scale-free network [24] with \(p=204\) (Fig. 1c) is analyzed as one of the real networks. Based on the information given in [24], this graph displays the major component of the network which was generated by proteome data of Saccharomyces cerevisiae. Figure 2e shows the sensitivity only on the connections between the hubs and other nodes under various penalty constants. In this figure, the sample size is fixed to \(N=20{,}400\) and the constant \(\rho \) changes. Our method is more robust under various constants \(\rho =\{0.001-0.05\}\). The histogram of all connections is shown in Fig. 2f. Table 1(iii) shows the results on this data set with a specific scenario. Figure 3 shows the results of the precision matrix estimation. This figure shows that the proposed algorithm protects the hubs. Figure 4 shows the partial section of this network. In this figure, the quality assessment on 4 hubs and their connections is shown. These results are obtained when the sensitivity measurements of all 204 variables’ connections are 93, 92 and 98 % for Covsel [16], Glasso [6] and the proposed method, respectively. The results reveal that the loss of connections between the hubs and other nodes is decreased by our algorithm when the sparsity is needed. The specificity rates are 98.8, 98.4 and 98.1 % for Covsel [16], Glasso [6] and the proposed method, respectively. As a second real data, we test the methods on a yeast data set [27, 28] with \(p=2361\). Estimation methods for the interaction detection have led to the discovery of thousands of interactions between proteins. This yeast data set consists of protein–protein interaction network described and analyzed in [27] and available as an example in the software package [28]. Figure 5a shows the sensitivity only on the connections between the hubs and other nodes under various sample sizes. Figure 5b shows the histogram of the number of connections with respect to the number of variables. Table 1(iv) shows the results on this data set with a specific scenario. Also, the execution time of each method is shown in this table. Figure 6 shows the partial section of this network. In this figure, the quality assessment on 3 hubs and their connections is shown. In this real network, we accept that three variables (“120,” “126” and “135”) are the hubs for this partial region. These results are obtained when the sensitivity measurements of all 2361 variables’ connections are 93, 92 and 98 % for Covsel [16], Glasso [6] and the proposed method, respectively. The specificity rates are 98.7, 99.7 and 99.3 % for Covsel [16], Glasso [6] and the proposed method, respectively. The false-positive (FP) connections are not shown since there are very few edges showing wrong dependencies for three methods. The results reveal that the loss of connections between the hubs and other nodes is decreased by our algorithm when the sparsity is needed.
A larger real data set [25] is a network that analyze the connections in Facebook. One of the objectives of analyzing the social data is to discover the connections and users’ social circles. The number of subjects or variables is 4039 in these data, whereas the average connection number is 21.88 per variable. The biggest hub in this data set has 1043 connections. We evaluate the performance of the proposed and alternative methods with different scenarios as shown in Table 2. The results show that hubs are more protected with our proposed method using various parameters, while the other measurements (i.e., sensitivity, specificity and MCC) are comparable with other alternatives.
Finally, we assess our proposed method on the Extraction, Visualization and Analysis (EVA) of corporate inter-relationships (US Corporate Ownership) data [26]. These data are constructed using the telecommunications and media industries with ownership networks with 6726 relationships among 8343 companies. The analysis of these data reveals a highly clustered network, with over 50 % of all companies connected to one another in a single one. In another words, a link (X, Y) revealing a connection from company X to company Y exists in the network if in the company X has the ownership of the company Y [29]. Also, the ownership activity is highly unbalanced such that 90 % of companies have no more than one relationship, whereas the top ten companies are parents for over 24 % of all relationships. These data can be a representative population for the current inter-relationships. The results and comparison with the alternatives are shown in Table 3. Again, in this experiment, we prove that our proposed idea is more robust, especially on the hubs when using various parameters.
4 Conclusions
This article has presented a new Gaussian graphical model for the scale-free and/or hub-dominated networks. We proposed a new convex log-likelihood formulation, which promotes the sparseness of the precision matrix and scale-free feature of graphical topology together. Our proposed method was assessed on two synthetic and four real networks which have various variable numbers. As can be seen from the experiments, the adaption of the our method to the various data types such as yeast, social and business interactions is very straightforward. The proposed method is more robust under various effects and is able to guard hubs; hence, scale-free networks are protected to the possible damages to hubs. There are several ways to extend this research. For example, the choice of the penalty constant for the hubs may be studied although our proposed method is not very sensitive to the penalty constant selection based on our experiments. The method appears to be a potentially valuable technique for analyzing various large-scale databases.
References
Aslan, M.S., Chen, X.W., Cheng, H.: Learning sparse and scale-free networks. In: Data Science and Advanced Analytics (DSAA), 2014 IEEE International Conference on 2014, pp. 326–332 (2014)
Dempster, A.P.: Covariance selection. Int. Biom. Sel. 28(1), 157–175 (1972)
Honorio, J., Samaras, D., Rish, I., Cecchi, G.: Variable selection for Gaussian graphical models. In: AISTATS, pp. 538–546 (2012)
Meinshausen, N., Buhlmann, P.: High dimensional graphs and variable selection with the lasso. Ann. Stat. 34, 1436–1462 (2006)
Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. 58(1), 267–288 (1996)
Friedman, J., Hastie, T., Tibshirani, R.: Sparse inverse covariance estimation with the graphical lasso. Biostatistics 9(3), 432–441 (2007)
Rothman, A., Levina, P., Zhu, J.: Sparse permutation invariant covariance estimation. Electron. J. Stat. 2, 494–515 (2008)
Fan, J., Feng, Y., Wu, Y.: Network exploration via the adaptive lasso and scad penalties. Ann. Appl. Stat. 3(2), 521–541 (2009)
Banerjee, O., Ghaoui, E.L., d’Aspremont, A., Natsoulis, G.: Model selection through sparse maximum likelihood estimation. J. Mach. Learn. Res. 9, 485–516 (2007)
Guo, J., Wang, S.: A constrained \(l1\) minimization approach to sparse precision matrix estimation. Am. Stat. Assoc. 106(494), 594–607 (2011)
Dahl, J., Vandenberghe, L., Roychowdhury, V.: Covariance selection for non-chordal graphs via chordal embedding. Methods Softw. 23(4), 485–516 (2008)
Yuan, M., Lin, Y.: Model selection and estimation in the gaussian graphical model. Biometrika 94(1), 19–35 (2007)
d’Aspremont, A., Banerjee, O., El Ghaoui, L.: First-order methods for sparse covariance selection. SIAM J Matrix Anal. Appl. 30(1), 56–66 (2008)
Lu, Z.: Adaptive first-order methods for general sparse inverse covariance selection. SIAM J. Matrix Anal. Appl. 31(4), 2000–2016 (2010)
Ravikumar, P., Wainwright, M.J., Raskutti, G., Yu, B.: High-dimensional covariance estimation by minimizing \(l1\)-penalized log-determinant divergence. Electron. J. Stat. 5, 935–980 (2011)
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 3(1), 1–122 (2010)
Barabási, A.L.: Scale-free networks: a decade and beyond. Science. 325(5939), 412 (2009)
Liu, Q., Ihler, A.T.: Learning scale free networks by reweighted l1 regularization. In: International Conference on Artificial Intelligence and Statistics, pp. 40–48 (2011)
Liu, C., Du, W.B., Wang, W.X.: Particle swarm optimization with scale-free interactions. PloS one 9(5), e97822 (2014)
Zhang, C., Huang, H.: Optimal control strategy for a novel computer virus propagation model on scale-free networks. Phys. A Stat. Mech. Appl. 451, 251–265 (2016)
Liu, H., Yin, R., Liu, B., Li, Y.: A scale-free topology model with fault-tolerance and intrusion-tolerance in wireless sensor networks. Comput. Electr, Eng (2016)
Glowinski, R., Marrocco, A.: Sur lapproximation, par elements finis dordre un, et la resolution, par penalisation-dualie, dune classe de problems de dirichlet non lineares. Rev. Franca. Automat. Inf. Rech. Oper. 9, 4176 (1975)
Gabay, D., Mercier, B.: A dual algorithm for the solution of nonlinear variational problems via finite element approximations. Comput. Math. Appl. 2, 1740 (1976)
Wuchty, S.: Scale-free behavior in protein domain networks. Mol. Biol. Evol 18, 1694–1702 (2001)
McAuley, J.J., Jure Leskovec.: Learning to discover social circles in ego networks. In: NIPS, vol. 2012, pp. 548–56 (2012)
Norlen, K., Lucas, G., Gebbie, M., Chuang, J.: EVA: Extraction, visualization and analysis of the telecommunications and media ownership network. In: Proceedings of International Telecommunications Society 14th Biennial Conference (ITS2002), Seoul Korea (2002)
Sun, S., Ling, L., Zhang, N., Li, G., Chen, R.: Topological structure analysis of the protein–protein interaction network in budding yeast. Nucleic Acids Res. 31(9), 2443–2450 (2003)
Software package protein interaction network (pin) at http://www.bioinfo.org.cn/PIN/
Acknowledgments
This work is supported by the US National Science Foundation award (IIA-1028098).
Author information
Authors and Affiliations
Corresponding author
Additional information
Invited and extended version of our paper [1] presented at the International Conference on Data Science and Advanced Analytics, Shanghai, China, October 30–November 1, 2014.
Rights and permissions
About this article
Cite this article
Aslan, M.S., Chen, XW. & Cheng, H. Analyzing and learning sparse and scale-free networks using Gaussian graphical models. Int J Data Sci Anal 1, 99–109 (2016). https://doi.org/10.1007/s41060-016-0009-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41060-016-0009-y