Soft configuration model

From HandWiki
Short description: Random graph model in applied mathematics

In applied mathematics, the soft configuration model (SCM) is a random graph model subject to the principle of maximum entropy under constraints on the expectation of the degree sequence of sampled graphs.[1] Whereas the configuration model (CM) uniformly samples random graphs of a specific degree sequence, the SCM only retains the specified degree sequence on average over all network realizations; in this sense the SCM has very relaxed constraints relative to those of the CM ("soft" rather than "sharp" constraints[2]). The SCM for graphs of size n has a nonzero probability of sampling any graph of size n, whereas the CM is restricted to only graphs having precisely the prescribed connectivity structure.

Model formulation

The SCM is a statistical ensemble of random graphs G having n vertices (n=|V(G)|) labeled {vj}j=1n=V(G), producing a probability distribution on 𝒢n (the set of graphs of size n). Imposed on the ensemble are n constraints, namely that the ensemble average of the degree kj of vertex vj is equal to a designated value k^j, for all vjV(G). The model is fully parameterized by its size n and expected degree sequence {k^j}j=1n. These constraints are both local (one constraint associated with each vertex) and soft (constraints on the ensemble average of certain observable quantities), and thus yields a canonical ensemble with an extensive number of constraints.[2] The conditions kj=k^j are imposed on the ensemble by the method of Lagrange multipliers (see Maximum-entropy random graph model).

Derivation of the probability distribution

The probability SCM(G) of the SCM producing a graph G is determined by maximizing the Gibbs entropy S[G] subject to constraints kj=k^j, j=1,,n and normalization G𝒢nSCM(G)=1. This amounts to optimizing the multi-constraint Lagrange function below:

(α,{ψj}j=1n)=G𝒢nSCM(G)logSCM(G)+α(1G𝒢nSCM(G))+j=1nψj(k^jG𝒢nSCM(G)kj(G)),

where α and {ψj}j=1n are the n+1 multipliers to be fixed by the n+1 constraints (normalization and the expected degree sequence). Setting to zero the derivative of the above with respect to SCM(G) for an arbitrary G𝒢n yields

0=(α,{ψj}j=1n)SCM(G)=logSCM(G)1αj=1nψjkj(G)  SCM(G)=1Zexp[j=1nψjkj(G)],

the constant Z:=eα+1=G𝒢nexp[j=1nψjkj(G)]=1i<jn(1+e(ψi+ψj))[3] being the partition function normalizing the distribution; the above exponential expression applies to all G𝒢n, and thus is the probability distribution. Hence we have an exponential family parameterized by {ψj}j=1n, which are related to the expected degree sequence {k^j}j=1n by the following equivalent expressions:

kq=G𝒢nkq(G)SCM(G)=logZψq=jq1eψq+ψj+1=k^q, q=1,,n.

References

  1. van der Hoorn, Pim; Gabor Lippner; Dmitri Krioukov (2017-10-10). "Sparse Maximum-Entropy Random Graphs with a Given Power-Law Degree Distribution". 
  2. 2.0 2.1 Garlaschelli, Diego; Frank den Hollander; Andrea Roccaverde (January 30, 2018). "Coviariance structure behind breaking of ensemble equivalence in random graphs". http://eprints.imtlucca.it/4040/1/1711.04273.pdf. 
  3. Park, Juyong; M.E.J. Newman (2004-05-25). "The statistical mechanics of networks".