1. Introduction
In this work we study the bilateral preference graphs
$\mathit{LK}(n,k)$
proposed by La and Kabkab (
$\textsf{LK}$
) [Reference La and Kabkab10], and prove their connectivity conjecture. These graphs are subgraphs of the complete graph
$K_n$
, and are constructed as follows. Assign independent and identically distributed (i.i.d.), uniform in [0, 1] weights to the
$\binom{n}{2}$
edges. Then, edge (i, j) is present in
$\mathit{LK}(n,k)$
if there is bilateral agreement in the preferences of vertices i and j for each other – if it is among the k edges of lowest weight incident to vertex i, and among the k edges of lowest weight incident to vertex j. Due to the bilateral agreement in the preferences, these graphs differ from both the Cooper–Frieze (
$\textsf{CF}$
)
$\mathit{CF}(n,k)$
random graphs [Reference Cooper and Frieze3] that are constructed via unilaterally preferred edges, and the (homogeneous or inhomogeneous) Erdős–Rényi (
$\textsf{ER}$
)
$\mathit{G}(n,p)$
random graphs [Reference Erdős and Rényi6, Reference Erdős and Rényi7, Reference Rényi13, Reference Shang14] where the edges are chosen independently.
Our main results, Theorems 3 and 6, can be collectively summarized as follows.
Theorem 1. The random graph
$\mathit{LK}(n,k)$
with
$k \leq \log (n) + t'\log \log (n)\sqrt{\log (n)}$
for
$t'<-\frac{1}{2}$
is not connected with high probability, and when
$k \geq \log (n) + t'\log \log (n)\sqrt{\log (n)}$
for
$t'>\frac{1}{2}$
, it is connected with high probability.
The proof of the disconnectedness result, Theorem 3, is established using the second moment method in which we prove that the expected number of isolated vertices grows to infinity and show that asymptotically any pair of vertices is isolated in an independent fashion. From this, we show that with high probability there exists an isolated vertex that yields the disconnectedness result. For the connectivity part, Theorem 6, we take the same approach as in [Reference La and Kabkab10], namely, find
$\textsf{ER}$
sub and super random graphs with n vertices of the
$\mathit{LK}(n,k)$
graph. This analysis then works by excluding the possibility of having components with r vertices for
$10\leq r\leq n/2$
. Independently, we also rule out the existence of components of size O(1). Note that the connectivity/disconnectedness results hold beyond the regime where we can develop a simple association with
$\textsf{ER}$
random graphs, as in [Reference La and Kabkab10]. While we don’t present the proof here, using the methodology of Cooper and Frieze [Reference Cooper and Frieze3], we get sharper connectivity/disconnectedness results in comparison to [Reference La and Kabkab10], but the final result is still not as sharp as Theorems 6 and 3.
In addition to the results on connectivity of
$\mathit{LK}(n,k)$
graphs, we also present results on the asymptotic behavior of the mean degree – see Theorems 7 and 8 – which we summarize below.
Theorem 2. For
$\mathit{LK}(n,k)$
graphs, suppose D represents the degree of a randomly chosen vertex. If
$k=o(\sqrt{n})$
, then we have the following asymptotic characterization for the mean degree of
$\mathit{LK}(n,k)$
graphs:
If we further assume that
$k=o(n^{1/3})$
and k grows to infinity as n goes to infinity, then the average degree has the asymptotic behavior
Remark 1. Theorems 3 and 6 in combination with Theorem 8 show that the
$\textsf{LK}$
random graph family provides another instance (except for a small gap around
$\log(n)$
) wherein, on the one hand, the mean degree being strictly greater than
$\log(n)$
implies connectivity, and, on the other hand, the mean degree being strictly smaller that
$\log(n)$
implies the existence of isolated vertices. For the
$\textsf{LK}$
family this holds despite the dependencies in the graph construction, unlike either the homogeneous or inhomogeneous
$\textsf{ER}$
random graph family. Finally, note also that the asymptotic characterization for the mean degree of
$\mathit{LK}(n,k)$
graphs spans a wider range of k when compared to the results of [Reference Moharrami, Subramanian, Liu and Sundaresan11] (to be discussed later). Note also that our precise characterization of the mean degree yields a correction to the affine behavior conjectured in [Reference La and Kabkab10, Section 6.4].
1.1. Open problems
Whereas Theorems 3 and 6 establish the conjecture by La and Kabkab, they still do not precisely identify the connectivity threshold in comparison to existing results on the (homogeneous)
$\textsf{ER}$
model [Reference Rényi13], or even the
$\textsf{CF}$
model [Reference Cooper and Frieze3]; specifically, if
$\log (n) -0.5\log \log (n)\sqrt{\log (n)} \leq k \leq \log (n) + 0.5\log \log (n)\sqrt{\log (n)}$
, then it is not known whether connectivity holds or not for the
$\textsf{LK}$
model. We believe this arises due to our use of analytically simpler sufficient conditions to prove our results – for example, to show that vertex i is isolated, we insist that all of its edges in
$\mathbb{K}_n$
are not preferred by its top k neighbors, with no intersection between the preferred edges of the neighbors. We use an elaborate sufficient condition to prove connectivity. Hence, a more elaborate (or even exact) set of sufficient conditions for both regimes could close the gap – this is an open problem for future work. Similarly, the proof methodology used to establish connectivity in [Reference La and Kabkab10] also allows La and Kabkab to determine the diameter of the
$\mathit{LK}(n,k)$
graphs, so another open problem is characterizing the diameter for the additional range of k(n) from Theorem 6 where connectivity holds. Finally, two additional questions are worth investigating. The first is to characterize the limiting distribution of
$(D-\mathbb{E}[D])/\sqrt{\mathrm{Var}[D]}$
. We expect this to be hard to answer owing to the complicated dependence structure of the graph. Note that even
$\mathrm{Var}[D]$
is hard to calculate – using our method in the proof of Theorem 8 to calculate it is hard due to the dependence of the connectivity of the nodes. The second is to find conditions under which Theorem 8 can be generalized to a concentration of node degrees around the mean degree, which is a result that holds for
$\textsf{ER}$
random graphs [Reference Draief and Massoulié4, Chapter 4] when
$p(n)={t\log(n)}/{n}$
(for
$t>0$
and all large enough n).
1.2. Organization of the paper
We start by discussing related work in Section 1.3. Thereafter, we formalize the mathematical model in Section 2. In Section 3 we show how the multinomial distribution arises in the limit of
$n\rightarrow\infty$
via a connection to an infinite urn model. Sections 4, 5, and 6 then establish the connectivity results and the characterization of the mean degree of the graph, respectively.
1.3. Related work
Graphs have been used to study interesting phenomena in many application domains [Reference Easley and Kleinberg5, Reference Jackson9, Reference Newman12]. They have been studied using probabilistic tools for about 50 years [Reference Draief and Massoulié4, Reference Van Der Hofstad15], starting with Gilbert in [Reference Gilbert8] and Erdős and Rényi in [Reference Erdős and Rényi6, Reference Erdős and Rényi7]. A (homogeneous) Erdős–Rényi random graph with n vertices is characterized by a parameter
$0\leq p(n) \leq 1$
that is the probability of existence of each potential undirected edge between two vertices. A realization of a graph from the
$\textsf{ER}$
model is denoted by
$\mathit{G}(n,p)$
, where the dependence of p on n is suppressed for brevity. Edges then appear independently based on Bernoulli coin tosses with probability p(n) for heads, and with no intrinsic preference on the edges on the part of the vertices. It has been shown [Reference Erdős and Rényi6, Reference Erdős and Rényi7] that if
$p(n)=t \log(n)/n$
, with high probability the graph
$\mathit{G}(n,p)$
is connected for
$t>1$
and disconnected for
$t<1$
; the refined result [Reference Rényi13] is that if (for large enough n)
$p(n)=(\log(n)+c)/n$
, then the graph is connected with probability
$\exp(-{\mathrm{e}}^{-c})$
. The critical property needed for connectivity to hold is that the mean degree
$t\log(n)$
is strictly greater than
$\log(n)$
, and for isolated vertices to exist is that the mean degree is strictly less than
$\log(n)$
; these results also generalize to inhomogeneous
$\textsf{ER}$
graphs – see [Reference Shang14].
Cooper and Frieze in [Reference Cooper and Frieze3] described a new family of random graphs based on preferences constructed using a distance measure. They considered the complete graph
$K_n$
and assigned independent uniform [0, 1] random variables as distances to all edges. They then kept the k(n) shortest edges incident to each vertex. In this model, the concept of distance induces a preference order on all edges: the shorter an edge, the higher its preference. Since the distances are i.i.d., the preference order on all the edges will be chosen uniformly over the set of all permutations. We say a vertex proposes an edge if the edge belongs to the set of the k(n) most-preferred edges incident to it. Note that in the Cooper–Frieze model, an edge is kept if at least one of its endpoints proposes it, which is a unilateral perspective. A realization of a graph from the
$\textsf{CF}$
model is denoted as
$\mathit{CF}(n,k)$
, with the dependence of k on n suppressed for brevity.
Preference relations of the sort used in the
$\textsf{CF}$
model create a complicated structure on any resulting graph. For example, keeping edges based on their ranking even in a unilateral manner induces an edge dependency in the
$\textsf{CF}$
model (which is not the case for the
$\textsf{ER}$
model). Then, [Reference Cooper and Frieze3] proved that the graph
$\mathit{CF}(n,k)$
is connected with high probability for
$k\geq 3$
, and also provided upper and lower bounds for the probability of connectivity as
$n\rightarrow\infty$
when
$k=2$
.
Bilateral preference random graphs, introduced in [Reference La and Kabkab10], are constructed using the same parameter k(n) as the
$\textsf{CF}$
random graphs. Considering vertices as agents, all vertices are assumed to have their own preferences on the potential edges with others via a priority or preference order over other vertices, where the individual vertex preferences result from a single global preference order on the
$\binom{n}{2}$
possible edges. Then, in the La–Kabkab model, in contrast to both the
$\textsf{ER}$
and the
$\textsf{CF}$
models, an edge is drawn if and only if each end vertex has the other vertex in its k(n) preferred vertices. In other words, an edge is formed if and only if both end vertices propose the edge to the other vertex, which then leads to the need for bilateral preference. Hence, the parameter k(n) is now the maximum number of other vertices each vertex wants as its neighbors. A graph realization from the
$\textsf{LK}$
model is denoted by
$\mathit{LK}(n,k)$
; the dependence of k on n is again suppressed for brevity.
Following [Reference La and Kabkab10], we can interpret the bilateral-preference-based
$\textsf{LK}$
model as a network formation process conducted via a game (selfish objective maximization) among bounded rational agents at the vertices. These kinds of network formation processes [Reference Easley and Kleinberg5, Reference Jackson9, Reference Newman12] have been studied in social science and economics. We assume that the n agents are aware of their own benefits from the potential pairwise connections with others, where the benefit is assessed via an appropriate cost or distance measure – we assume, without loss of generality, that the pairwise costs or distances generated are independent and either uniformly distributed in [0, 1] or exponentially distributed with parameter 1. We point the reader to [Reference La and Kabkab10, Observation O-3] for the reason why either of these choices of distribution (or any other continuous distribution) results in the same realization of graphs. Again following [Reference La and Kabkab10], the agents are bounded rational, and only use local information instead of global information. Finally, the agents also have limited memory, so that each agent prefers the k(n) most profitable/valuable connections based solely on the cost or distance of the potential edge instead of some global objective like connectivity. Then, each edge (connection) will exist if and only if both parties prefer it, and from this process the graph of pairwise connections will be generated.
A core question studied in [Reference La and Kabkab10] is to determine the k(n) that results in the graphs produced being connected. It showed that the following results hold with high probability: (i) if
$k(n) > C \log(n)$
for
$C=2.4625$
, then the graph is connected; and (ii) if
$k(n)< c \log(n)$
for
$c=0.5$
, then the graph has isolated vertices and so is not connected. Furthermore, using extensive simulations, La and Kabkab also conjectured that the connectivity threshold was exactly
$k(n)=\log(n)$
, which is, surprisingly, the same as the threshold for connectivity of the
$\textsf{ER}$
random graph family in terms of the mean degree. Note once again that k(n) is the maximum degree for an
$\mathit{LK}(n,k)$
graph. As discussed earlier, we establish this conjecture by providing a finer characterization of when connectivity holds and when it does not.
A different line of work from the connectivity question for
$\mathit{LK}(n,k)$
graphs is that of [Reference Moharrami, Subramanian, Liu and Sundaresan11], which introduced a new branching process called the Erlang weighted tree (EWT) as the local weak limit of the
$\textsf{LK}$
model graphs in the sparse regime. Specifically, in [Reference Moharrami, Subramanian, Liu and Sundaresan11] the parameter k(n) is an (finite) random parameter k for each vertex that is independently chosen and identically distributed with a distribution on
$\mathbb{N}$
with finite mean. In this regime,
$\textsf{LK}$
graphs are shown to be different from
$\textsf{ER}$
graphs. This was shown by finding the degree distribution of the root vertex and also its mean degree, which coincide with the asymptotic degree distribution and mean degree for
$\textsf{LK}$
graphs, respectively. As discussed earlier, in Section 6 we study the asymptotics of the mean degree of an
$\mathit{LK}(n,k)$
graph for a wider range of k, and as a consequence of results we present an alternate derivation of the mean degree from [Reference Moharrami, Subramanian, Liu and Sundaresan11] when k is deterministically chosen and finite. [Reference Moharrami, Subramanian, Liu and Sundaresan11] also discussed the probability of extinction of an EWT, and conjectured its relevance to the size of the giant component [Reference Draief and Massoulié4, Reference Erdős and Rényi7, Reference Van Der Hofstad15] of an
$\textsf{LK}$
graph, when there is one.
2. Mathematical model
Consider the complete undirected graph
$\mathbb{K}_n$
with vertices
$[n]=\{1,2,\ldots,n\}$
and edges
$\{(i,j)\colon 1\leq i,j\leq n, i\neq j\}$
, where we assume
$(i,j)=(j,i)$
. Let k(n) be an integer such that
$1\leq k(n)\leq n$
; henceforth, to avoid cumbersome notation, we will use k instead of k(n). We assign i.i.d. random variables called priority scores to all edges of
$K_n$
. For any explicit calculations, we can assume they are uniformly distributed in [0, 1] or exponentially distributed with parameter 1; this holds because we will only be interested in the order statistics. We denote the score of edge (i, j) by
$V(i,j)=V(j,i)$
. The set of all scores of the edges of vertex
$i\in[n]$
is denoted by
$\mathcal{V}_i=\{V(i,1),\ldots,V(i,i-1), V(i,i+1),\ldots, V(i,n)\}$
, and all the associated edges by
$\mathcal{E}_i=\{(i,1),\ldots, (i,i-1),(i,i+1),\ldots,(i,n)\}$
. Without loss of generality, we can also assume that the scores are distinct; then
$(R_i^j)_{1\leq j\leq n-1}$
represents an order on
$[n]\setminus\{i\}=\{1,2,\ldots,i-1,i+1,\ldots,n\}$
based on V(i, j) values. In other words, for each
$1\leq i\leq n$
, the random vector
$\mathcal{R}_i=(R_i^1,R_i^2,\ldots,R_i^{n-1})$
is a permutation of
$[n]\setminus\{i\}$
in which
$V(i,R_i^1)> V(i,R_i^2)>\cdots> V(i,R_i^{n-1})$
. As the scores are chosen i.i.d. from a continuous distribution, the distribution of the random vector
$\mathcal{R}_i$
is uniform among all permutations of
$[n]\setminus\{i\}$
as it only depends on the order statistics. The scores also impose a permutation over all the edges. This plays an important role in defining the bilateral preference
$\textsf{LK}$
random graphs.
Let V(i, j) be realized for all edges (i, j), and parameter k be fixed. Then we can determine two different classes of random graphs on vertices [n]. We first define the notion of preference. If V(i, j) is among the k largest scores in
$\mathcal{V}_i$
, i.e.
$j\in \mathcal{R}_i^{\leq k}:=\{R_i^1,R_i^2,\ldots,R_i^k\}$
or
$(i,j)\in\mathcal{E}_i^{\leq k} := \{(i,R_i^1),\ldots,(i,R_i^k)\}$
, we say that vertex i proposes/prefers edge (i, j). Moreover,
$\mathcal{E}_i:=\mathcal{E}_i^{\leq n-1}$
denotes all the edges incident to vertex i. The first model introduced in [Reference Cooper and Frieze3] let the edge (i, j) be present if at least one of i or j prefers (i, j). Remember that we denote realizations of these random graphs by
$\mathit{CF}(n,k)$
, which we call the class of unilaterally proposed graphs. The second model described in [Reference La and Kabkab10] requires a preference by both vertices for an edge to appear, which we call bilateral preference. Again, we denote realizations of this class of graph by
$\mathit{LK}(n,k)$
, and we call it the class of bilateral preference random graphs. We can also construct
$\textsf{ER}$
random graphs through independent node preferences: each node prefers the possible
$n-1$
edges independently with probability
$\sqrt{p(n)}$
with an edge forming only via bilateral preference.
Both
$\mathit{CF}(n,k)$
and
$\mathit{LK}(n,k)$
graphs only depend on the order of V(i, j), and not the precise values. This is the consequence of the scores (costs or distances) being i.i.d. from a continuous distribution, which results in a uniformly drawn permutation among all permutations of edges of
$\mathbb{K}_n$
; again, note [Reference La and Kabkab10, Observation O-3]. Assigning a random variable to each edge only helps to motivate the graph construction, but while determining any underlying probability we will typically use the random permutations on edges viewpoint.
3. Asymptotically equivalent distribution
In our analysis we will use the abstraction of an infinite urn model. The following two lemmas enable us to connect the probability of an event in the finite permutations space to an event in the infinite urn model. In essence, we will show that the appropriate probabilities converge to a binomial distribution in Lemma 1 and to a negative multinomial distribution in Lemma 2. Before delving into details, we will consider the set of all permutations with order restrictions on some elements. The notation
$x\succ y$
denotes that x appears earlier than y in the permutations. These simple orders can be combined using and and or to create more complex order restrictions on the permutation. For instance, when we refer to the set of permutations of
$\{1,2,3\}$
with order restrictions
$1\succ 2\ $
and
$\ 1\succ 3$
, this narrows down the set of all
$3!=6$
permutations to just
$\{(1,2,3),(1,3,2)\}$
. If we instead use the restriction
$1\succ 2$
or
$1\succ 3$
, the resulting set of permutations is
$\{(1,2,3),(1,3,2),(2,1,3),(3,1,2)\}$
.
Lemma 1. Assume we have
$M=m_0+m_1+\cdots +m_s$
objects, each assigned a type, where
$m_i$
denotes the number of objects of type i for each
$0\leq i \leq s$
. Consider a uniformly random permutation of all M objects with some order restrictions which are only for objects of type 0. Let X represent the number of type 0 objects within the first l positions of this permutation, where
$l<M$
. The law of X for a given
$j<l$
is given by
\begin{equation} \mathbb{P}(X=j) = \binom{l}{j}\Bigg(\prod_{i=0}^{j-1}(m_0-i)\Bigg)\Bigg(\prod_{i=0}^{l-j-1} (M-m_0-i)\Bigg) \frac{(M-l)!}{M!}. \end{equation}
We further assume there exists a parameter n such that
$s,l=o(n^{1/4})$
, and there are constants
$h_i>0$
with
$m_i=nh_i+\varepsilon_i$
where
$\sum_{i=1}^s|\varepsilon_i|=o(n^{1/4})$
and
$\max_{1\leq i\leq s}h_{i}=O(1)$
. Under these conditions, the asymptotic probability is given by
as
$n\rightarrow\infty$
.
Proof. To compute this probability we will count the number of permutations with j elements of type 0 at the first l place then divide it by the number of all permutations. Any given order restrictions on objects of type 0 result in both the numerator and denominator of this fraction being divided by the number of symmetries. As a result, we can assume there is no order restriction on objects of type 0.
The first part is straightforward by choosing those j places at the first l observations for objects of type 0, and
$m_0 - j$
places in the remaining part. Next, by counting the number of desired arrangements for objects of type 0 and other types we arrive at the formula
Now (1) can be derived through straightforward calculations. For the second part, we first simplify
${(M-l)!}/{M!}$
to obtain
\begin{align*}\binom{l}{j}\Bigg(\frac{\prod_{i=0}^{j-1}(m_0-i)\prod_{i=0}^{l-j-1}(M-m_0-i)}{\prod_{i=0}^{l-1}(M-i)}\Bigg).\end{align*}
Note that
$m_0-i = h_0 n+o(n^{1/4})$
since
$i\leq j < l = o(n^{1/4})$
. Similarly,
$M-m_0-i=$
$(h_1+\cdots +h_s)n+o(n^{1/4})\,M-m_0-i=(h_1+\cdots +h_s)n+o(n^{1/4})$
and
$M-i=(h_0+h_1+\cdots+h_s)n+o(n^{1/4})$
for each i. Substituting these expressions in the second term in the product above results in
\begin{align*} \frac{\big(h_0 n + o(n^{1/4})\big)^j\big((h_1+\cdots +h_s)n+ o(n^{1/4})\big)^{l-j}} {\big((h_0+h_1+\cdots +h_s)n+ o(n^{1/4})\big)^{l}}. \end{align*}
To derive (2), we factor out n from each term within the parentheses and simplify further, using the fact that
$j,l-j=o(n^{1/4})$
:
\begin{align*} & \frac{\big(h_0 + o(n^{-3/4})\big)^j\big(h_1+\cdots +h_s+ o(n^{-3/4})\big)^{l-j}} {\big(h_0+h_1+\cdots +h_s+ o(n^{-3/4})\big)^{l}} \\[6pt] & = \bigg(\frac{h_0^j(h_1+\cdots+h_s)^{l-j}}{(h_0+h_1+\cdots+h_s)^l}\bigg) \frac{((1 + {o(n^{-3/4})})/{h_0})^j(1+({o(n^{-3/4})}/({h_1+\cdots+h_s})))^{l-j}} {(1+({o(n^{-3/4})}/({h_0+h_1+\cdots+h_s})))^{l}} \\[6pt] & = \bigg(\frac{h_0^j(h_1+\cdots+h_s)^{l-j}}{(h_0+h_1+\cdots+h_s)^l}\bigg)\big(1+o(n^{-3/4})\big)^{2l} \\[6pt] & = \bigg(\frac{h_0^j(h_1+\cdots+h_s)^{l-j}}{(h_0+h_1+\cdots+h_s)^l}\bigg)\big(1+o(n^{-3/4})\big)^{o(n^{1/4})} = \bigg(\frac{h_0^j(h_1+\cdots+h_s)^{l-j}}{(h_0+h_1+\cdots+h_s)^l}\bigg)\big(1+o(n^{-1/2})\big). \end{align*}
This completes the proof.
Lemma 2. Under the same assumptions as Lemma 1, consider M objects, each assigned a type from 0 to s, with
$m_i$
objects of type i for each
$0\leq i\leq s$
. Let
$X=(i_1,\ldots,i_s)$
be a random vector where
$i_j$
represents the number of occurrences of objects of type j before the first occurrence of objects of type 0 for
$1\leq j\leq s$
in a uniformly random permutation. Then the law for X is given by
Similar to Lemma 1, if we further assume
$s=o(n^{1/4})$
and
$i_j < m_j=nh_j+\varepsilon_j$
with
$\sum_{j=1}^s|\varepsilon_j|, \sum_{j=1}^s i_j=o(n^{1/4})$
, then the probability (3) asymptotically approaches
as
$n\rightarrow\infty$
.
Proof. It is easy to check that the number of permutations of M objects with exactly
$i_j$
objects of type j at the first
$i_1+i_2+\cdots+i_s$
places for each
$1\leq j\leq s$
, and with the first object of type 0 at the
$(i_1+\cdots+i_s+1)$
th place, is
Dividing by the total number of permutations, i.e.
$M!$
, yields (3).
Then, note that
Moreover, using
$s=o(n^{1/4})$
,
$\sum_{j=1}^s|\varepsilon_j|=o(n^{1/4})$
, and
$\sum_{j=1}^s|i_j|=o(n^{1/4})$
, we have
\begin{align*} & \frac{(M-i_1-\cdots-i_s-1)!}{M!} \\[3pt] & \qquad = \frac{1}{M(M-1)\cdots(M-i_1-\cdots i_s)} \\[3pt] & \qquad = \frac{1}{\big((h_0+h_1+\cdots +h_s)n + o(n^{1/4})\big)^{i_1+\cdots+i_s+1}} \\[3pt] & \qquad = \frac{1}{((h_0+\cdots+h_s)n)^{i_1+\cdots+i_s+1}} \cdot \frac{1}{\big(1+({o(n^{-3/4})}/{(h_0+\cdots+h_s)})\big)^{i_1+\dots +i_s+1}} \\[3pt] & \qquad = \frac{1}{(h_0+\cdots+h_s)^{i_1+\cdots+i_s+1}n^{i_1+\cdots+i_s+1}} \cdot \frac{1}{(1+o(n^{-3/4}))^{i_1+\dots +i_s+1}}. \end{align*}
We substitute the last two asymptotic expressions along with
$m_0=h_0n(1+o(n^{-3/4}))$
into (3) to estimate the probability of the event in question:
\begin{align*} \mathbb{P}(X=(i_1,\ldots,i_s)) & = \bigg(\frac{h_0 h_1^{i_1}\cdots h_s^{i_s}(i_1+\cdots+i_s)!} {(h_0+h_1+\cdots+h_s)^{i_1+\cdots+i_s+1}i_1!\cdots i_s!}\bigg) \frac{\big(1+o(n^{-3/4})\big)^{i_1+\cdots+i_s+1}}{\big(1+o(n^{-3/4})\big)^{i_1+\cdots+i_s+1}} \\[4pt] & = \bigg(\frac{h_0 h_1^{i_1}\cdots h_s^{i_s}(i_1+\cdots+i_s)!} {(h_0+h_1+\cdots+h_s)^{i_1+\cdots+i_s+1}i_1!\cdots i_s!}\bigg)\big(1 + o(n^{-3/4})\big)^{2(i_1+\cdots+i_s+1)} \\[4pt] & = \bigg(\frac{h_0 h_1^{i_1}\cdots h_s^{i_s}(i_1+\cdots+i_s)!} {(h_0+h_1+\cdots+h_s)^{i_1+\cdots+i_s+1}i_1!\cdots i_s!}\bigg)\big(1 + o(n^{-3/4})\big)^{o(n^{1/4})} \\[4pt] & = \bigg(\frac{h_0 h_1^{i_1}\cdots h_s^{i_s}(i_1+\cdots+i_s)!} {(h_0+h_1+\cdots+h_s)^{i_1+\cdots+i_s+1}i_1!\cdots i_s!}\bigg)\big(1 + o(n^{-1/2})\big). \end{align*}
This completes the proof.
Lemmas 1 and 2 relate certain statistics in the space of uniformly random permutations of a large set of objects, each assigned a specific type, to the binomial and negative multinomial distributions. Specifically, under certain assumptions on the numbers of types and objects of each type, Lemma 1 establishes a relationship between the probability distribution of the number of occurrences of a particular type in the first l positions of this permutation (for sufficiently small l) and a binomial distribution. Additionally, Lemma 2 links the number of occurrences of non-zero types before the first occurrence of type 0 to a negative multinomial distribution. As mentioned earlier, this distributional convergence will help us greatly in our analysis. The special case required in this paper fixes all
$h_i\equiv1$
. Therefore, we state the following lemma.
Lemma 3. Under the same assumptions as Lemma 1, there are M objects, each assigned a type from 0 to s, with
$m_i$
objects of type i for
$0\leq i\leq s$
. We similarly assume the existence of a parameter n with
$s=o(n^{1/4})$
,
$m_i=n+\varepsilon_i$
so that
$\sum_{j=1}^s|\varepsilon_j|=o(n^{1/4})$
. Additionally, let
$l=o(n^{1/4})$
grow to infinity when
$n\rightarrow\infty$
under the extra condition that
$s=o(l)$
.
Denoting by X the number of objects of type 0 within the first l positions of this permutation, we have the following bound for its lower tail:
where
$t\leq l/s$
. In particular, if there are parameters k, a, b with
$k=o(n^{1/6})$
growing to infinity as
$n\rightarrow\infty$
,
$a,b=O(1)$
, and
$s,t=\sqrt{k}+O(1)$
so that
$l=(s-b)(k-s-a)$
, for sufficiently large n, then we have
$\mathbb{P}(X\leq t) \leq \exp\big({-}k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\big)$
.
Proof. We can assume
$k >s$
for sufficiently large n. It follows from
$t\leq l/s$
that
$\binom{l}{j}{s^{l-j}}/{(s+1)^l}$
is increasing for
$0\leq j\leq t$
. Using this fact and Lemma 1 with
$h_i=1$
, we have
\begin{align*} \mathbb{P}(X \leq t) & = \sum_{j=0}^{t}\binom{l}{j}\frac{s^{l-j}}{(s+1)^l}\big(1+o(n^{-1/2})\big) \\[3pt] & \leq (t+1)\binom{l}{t}\frac{s^{l-t}}{(s+1)^l}\big(1+o(n^{-1/2})\big) \\[3pt] & = \frac{t+1}{s^{t}}\binom{l}{t}\bigg(1-\frac{1}{s+1}\bigg)^l\big(1+o(n^{-1/2})\big) \\[3pt] & = \frac{t+1}{s^{t}}\binom{l}{t}\bigg(\bigg(1-\frac{1}{s+1}\bigg)^{s+1}\bigg)^{l/(s+1)}\big(1+o(n^{-1/2})\big) \\[3pt] & \leq \frac{t+1}{s^{t}}\binom{l}{t}\exp\bigg(\frac{-l}{s+1}\bigg)\big(1+o(n^{-1/2})\big) \\[3pt] & \leq \frac{t+1}{s^{t}} \cdot \frac{l^t}{t!}\exp\bigg(\frac{-l}{s+1}\bigg)\big(1+o(n^{-1/2})\big). \end{align*}
We can use the bound from Stirling’s approximation, i.e.
$t!>(t/{\mathrm{e}})^t$
, to get
\begin{align*} \mathbb{P}(X \leq t) & \leq \frac{(t+1){\mathrm{e}}^t l^t}{s^{t}t^t}\exp\bigg(\frac{-l}{s+1}\bigg)\big(1+o(n^{-1/2})\big) \\ & = \exp\bigg(\frac{-l}{s+1} + t\log\bigg(\frac{l}{ts}\bigg) + t + \log(t+1)\bigg)\big(1+o(n^{-1/2})\big). \end{align*}
By substituting
$s,t=\sqrt{k}+O(1)$
and
$l=(s-b)(k-s-a)=k\sqrt{k}+O(k)$
, we obtain
Additionally,
\begin{align*} t\log\bigg(\frac{l}{ts}\bigg) & = \left(\sqrt{k}+O(1)\right)\log\bigg(\frac{k\sqrt{k}+O(k)}{k + O(\sqrt{k})}\bigg) \\ & = \left(\sqrt{k}+O(1)\right)\log\left(\sqrt{k}+O(1)\right) = \frac{1}{2}\sqrt{k}\log(k) + O\left(\sqrt{k}\right). \end{align*}
This completes the proof.
4. Disconnectedness of
$\mathit{LK}(\mathit{n},\mathit{k})$
graphs for
$\boldsymbol{t} < \textbf{1}$
This section begins by recalling the negative multinomial distribution. In the special case that we need here, we present an urn model to interpret the distribution. Consider a sequence of i.i.d. random variables
$(X_i)_{i=1}^{\infty}$
, each taking
$k+1$
possible values
$\{0,1,\ldots,k\}$
uniformly at random. We refer to these outcomes as the types of objects, resulting in
$k+1$
distinct types. This sequence continues until the first occurrence of an object of type 0. The probability of observing
$i_1$
objects of type 1,
$i_2$
objects of type 2, …,
$i_k$
objects of type k before stopping follows the negative multinomial distribution
where
$i_j\geq 0$
for
$1\leq j\leq k$
. We now introduce some lemmas in order to bound certain probabilities that will appear in the proof.
Lemma 4. Let k,n be positive integers. With the definition in (4), we have
\begin{align*} \sum_{\substack{i_1+i_2+\cdots+i_k=n-1\\ \text{for all}\,j: i_j\geq 0}} f(i_1,i_2,\ldots,i_k) = \frac{1}{k+1}\bigg(\frac{k}{k+1}\bigg)^{n-1}. \end{align*}
Proof. In the urn model under consideration, i.e.
$(X_i)_{i=1}^{\infty}$
, the first appearance of 0 follows a geometric random variable with parameter
${1}/({k+1})$
. Therefore, the right-hand side is the probability that this random variable is n. The left-hand side sums over the probabilities of all possible numbers of other types observed prior to the nth draw.
Lemma 5. Let k be a positive integer and let
$0<\delta<1$
. Then
\begin{align*} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\ \text{for all}\,j:i_j\geq 0}} f(i_1,i_2,\ldots,i_k) \geq \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k\\ \text{for all}\,j:i_j\geq 0}} f(i_1,i_2,\ldots,i_k) - {\mathrm{e}}^{-4(k+1)}. \end{align*}
Proof. We start by bounding the term below, which we denote by r. We have
\begin{align*} r & = \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k\\[3pt] \text{for all}\,j:i_j\geq 0}}f(i_1,i_2,\ldots,i_k) - \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k \leq 4(k+1)^2\\[3pt] \text{for all}\,j:i_j\geq 0}}f(i_1,i_2,\ldots,i_k) \\[3pt] & = \sum_{\substack{4(k+1)^2 < i_1+i_2+\cdots+i_k\\[3pt] \text{for all}\,j:i_j\geq 0}}f(i_1,i_2,\ldots,i_k) \\[3pt] & = \sum_{n=4(k+1)^2}^{\infty}\frac{1}{k+1}\bigg(\frac{k}{k+1}\bigg)^n \qquad\qquad \mathrm{(Using\ Lemma \, 4)} \\[3pt] & = \bigg(\frac{k}{k+1}\bigg)^{4(k+1)^2} = \bigg(\bigg(1 - \frac{1}{k+1}\bigg)^{(k+1)}\bigg)^{4(k+1)} \leq {\mathrm{e}}^{-4(k+1)}, \end{align*}
and the result follows.
Lemma 6. Let k be a positive integer and
$0<\delta<1$
. Then
\begin{multline*} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\ \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots, i_k) \\ \geq (1 - k{\mathrm{e}}^{{-k\delta^2}/{5}})\sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\ \text{for all}\,j,i_j\geq 0}} f(i_1,i_2,\ldots,i_k). \end{multline*}
Before proving the result we point out the subtle difference in the summations on both sides of the last inequality: the left-hand side considers
$i_j \geq k$
, and the right-hand side considers
$i_j \geq 0$
, so the right-hand side summation is larger than the left. Hence, the result is non-trivial.
Proof. Let N denote the step in which we first observe a type 0 object in the sequence
$(X_i)_{i=1}^{\infty}$
. In other words,
$X_N=0$
and, for all
$i< N$
,
$X_i\neq 0$
. Moreover, we let
$I_j$
be the number of occurrences of objects of type j before step N. At each step prior to the Nth observation, whether an outcome is of type m or not, i.e.
$X_j=m$
, has a Bernoulli distribution with parameter
${1}/{k}$
. Denote this random variable by
$X_j^{(m)}$
, where j refers to the step number. For an arbitrary
$1\leq m \leq k$
, we have
$\mathbb{E}[X_j^{(m)}\mid j < N]={1}/{k}$
and
$\mathrm{Var}[X_j^{(m)}\mid j < N]=({k-1})/{k^2}$
. If we assume
$N=n>k^2(1+\delta)$
, then the assumption
$I_m<k$
implies
However, the expected value of the time average above given
$N=n$
is
${1}/{k}$
. Hence, the difference
indicates a lower tail event. To analyze this event, we apply the Bernstein inequality [Reference Boucheron, Lugosi and Massart2, Equation (2.10)] for
$n>k^2(1+\delta)$
:
\begin{align*} \mathbb{P}(I_m < k\mid N=n) & \leq \mathbb{P}\bigg(\frac{X_1^{(m)}+\cdots+X_{n-1}^{(m)}}{n-1} < \frac{1}{k(1+\delta)} \mid N=n\bigg) \\[8pt] & \leq \exp\bigg(\frac{-(n-1)\times{\delta^2}/({k^2(1+\delta)^2})} {{2(k-1)}/{k^2}+{2\delta}/({3k(\delta+1)})}\bigg) \\[8pt] & \leq \exp\bigg(\frac{-k^2(1+\delta)\times{\delta^2}/({k^2(1+\delta)^2})} {{2(k-1)}/{k^2}+{2\delta}/({3k(\delta+1)})}\bigg) \\[8pt] & = \exp\bigg(\frac{-3k^2\delta^2}{6(k-1)(\delta+1) + 2\delta k}\bigg) \leq \exp\bigg(\frac{-k\delta^2}{5}\bigg). \end{align*}
Using the union bound yields
$\mathbb{P}(\mathrm{for\ all}\ j\colon I_j\geq k\mid N=n) \geq 1 - k\exp({-k\delta^2}/{5})$
for
$n>k^2(1+\delta)$
. Therefore, we can apply this bound and definition (4) to obtain
\begin{align*} & \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\[3pt] \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots, i_k) \\[3pt] & =\sum_{k^2(1+\delta)< n < 4(k+1)^2}\;\sum_{\substack{\sum i_m=n-1,\\[3pt] \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots,i_k) \\[3pt] & = \sum_{k^2(1+\delta)< n < 4(k+1)^2}\mathbb{P}\,\,(\text{for all}\ j\colon I_j\geq k, N=n) \\[3pt] & = \sum_{k^2(1+\delta)< n < 4(k+1)^2}\mathbb{P}(N=n)\mathbb{P}\,\,(\text{for all}\ j\colon I_j\geq k\mid N=n) \\[3pt] & \geq \bigg(1 - k\exp\bigg(\frac{-k\delta^2}{5}\bigg)\bigg) \sum_{k^2(1+\delta)< n < 4(k+1)^2}\mathbb{P}(N=n) \\[3pt] & = \bigg(1 - k\exp\bigg(\frac{-k\delta^2}{5}\bigg)\bigg) \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\[3pt] \text{for all}\,j,i_j\geq 0}}f(i_1,i_2,\ldots,i_k). \end{align*}
This completes the proof.
Lemma 7. Let k be a positive integer, and let
$0\leq \delta\leq 1$
depend on k such that for sufficiently large k we have
$\delta\geq \sqrt{{6\log(k)}/{k}}$
. Then, the following asymptotic relationship holds:
\begin{equation} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\ \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots,i_k) = \exp\bigg(\bigg({-}k + \frac{1}{2}\bigg)(1+\delta) + o(1)\bigg) \end{equation}
as
$k\rightarrow\infty$
.
Proof. First, note that Lemma 4 and the second-order approximation
imply
\begin{align*} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k\\[4pt] \text{for all}\,j,i_j\geq 0}}f(i_1,i_2,\ldots,i_k) & = \sum_{N=k^2(1+\delta)+1}^{\infty}\frac{1}{k+1}\bigg(\frac{k}{k+1}\bigg)^{N-1} \\[4pt] & = \bigg(1 - \frac{1}{k+1}\bigg)^{k^2(1+\delta)} \\[4pt] & = \exp\bigg(\bigg({-}\frac{1}{k+1}-\frac{1}{2(k+1)^2}+\Theta\bigg(\frac{1}{k^3}\bigg)\bigg) k^2(1+\delta)\bigg) \end{align*}
\begin{align} & = \exp\bigg(\bigg({-}k + \frac{1}{2} + \Theta\bigg(\frac{1}{k}\bigg)\bigg)(1 + \delta)\bigg)\nonumber\\[4pt] & = \exp\bigg((-k + 0.5)(1+\delta) + \Theta\bigg(\frac{1}{k}\bigg)\bigg). \end{align}
We can now simply upper bound the left-hand side of (5) using (6):
\begin{align} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\nonumber\\ \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots,i_k) & \leq \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k\nonumber\\ \text{for all}\,j,i_j\geq 0}}f(i_1,i_2,\ldots,i_k) \nonumber\\ & = \exp\bigg((-k + 0.5)(1+\delta) + \Theta\bigg(\frac{1}{k}\bigg)\bigg) \nonumber\\ & = \exp\bigg((-k + 0.5)(1+\delta) + o(1)\bigg). \end{align}
We next use Lemmas 6 and 5 along with (6) to lower bound the left-hand side:
\begin{align*} & \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\[4pt] \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots,i_k) \\[4pt] & \qquad\qquad \geq (1 - k{\mathrm{e}}^{{-k\delta^2}/{5}})\sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\[4pt] \text{for all}\,j,i_j\geq 0}} f(i_1,i_2,\ldots,i_k) \\[4pt] & \qquad\qquad \geq (1 - k{\mathrm{e}}^{{-k\delta^2}/{5}}) \left(\sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k\\[4pt] \text{for all}\,j,i_j\geq 0}}f(i_1,i_2,\ldots, i_k) - {\mathrm{e}}^{-4(k+1)}\right) \\[4pt] & \qquad\qquad = \big(1 - k{\mathrm{e}}^{{-k\delta^2}/{5}}\big)\big({\mathrm{e}}^{(-k + {1}/{2})(1+\delta) + \Theta({1}/{k})} - {\mathrm{e}}^{-4(k+1)}\big) \\[4pt] & \qquad\qquad = {\mathrm{e}}^{(-k + {1}/{2})(1+\delta) + \Theta({1}/{k})}\big(1 - k{\mathrm{e}}^{{-k\delta^2}/{5}}\big) \big(1 - {\mathrm{e}}^{(\delta-3)k+(({-7+\delta})/{2})+\Theta({1}/{k})}\big). \end{align*}
The bounds
$\sqrt{{6\log k}/{k}}\leq \delta \leq 1$
yield
$k{\mathrm{e}}^{{-k\delta^2}/{5}}\leq {1}/{\sqrt[5]{k}}\rightarrow 0$
, and
Therefore, we can replace
$(1 - k{\mathrm{e}}^{{-k\delta^2}/{5}})$
and
$(1 - {\mathrm{e}}^{(\delta-3)k+(({-7+\delta})/{2})+\Theta({1}/{k})})$
by
${\mathrm{e}}^{o(1)}$
to deduce the following lower bound:
\begin{equation*} \sum_{\substack{k^2(1+\delta) < i_1+i_2+\cdots+i_k < 4(k+1)^2\\ \text{for all}\,j,i_j\geq k}}f(i_1,i_2,\ldots,i_k) \geq \exp((-k + 0.5)(1+\delta) + o(1)). \end{equation*}
This and (7) complete the proof.
We are now in a position to relate the lemmas above to the disconnectedness result. Assume
$I_j$
represents the event of vertex j being isolated. We will show that with high probability there exists a j such that
$I_j=1$
.
Lemma 8. For a graph of class
$\mathit{LK}(n,k)$
when
$k=o(\sqrt[8]{n})$
and
$k\rightarrow\infty$
, we have
for any
$\delta=\delta(k)$
with
$\sqrt{{6\log(k)}/{k}}\leq \delta(k) <1$
.
Proof. As mentioned earlier, any realization of independent random variables
$\{V(i,j)\}_{i,j}$
can be regarded as yielding a permutation of all edges
$\mathcal{E}$
. Without loss of generality, we can assume that vertex labels are sorted according to the scores of edges incident to vertex 1. In other words,
$V(1,2)>V(1,3)>\cdots>V(1,n)$
, or equivalently
$R_1^j=j+1$
for
$1\leq j\leq n-1$
. Then, for vertex 1 to be isolated, it is sufficient that, for each
$j\in\{2,3,\ldots,k+1\}$
, there exist at least k elements of
$\mathcal{E}_j$
(except (1, j)) appearing before all elements of
$\mathcal{E}_1$
in the permutation. Strictly speaking, if k elements of
$\mathcal{E}_j$
appear before
$(1,j+1)$
, then that would be necessary and sufficient. However, we consider a stricter condition insisting that k elements of
$\mathcal{E}_j$
appear before all elements of
$\mathcal{E}_1$
. This ensures that vertex j does not agree on edges (1, j) for
$2\leq j \leq k+1$
, and vertex 1 similarly does not agree upon edges (1, j) for
$k+2 \leq j \leq n$
, thereby isolating vertex 1.
To make this condition even stronger, we insist on no intersections. Define, for each
$2\leq i\leq k+1$
,
$\mathcal{E'}_i=\mathcal{E}_i \setminus \{(i,1),\ldots,(i,k+1)\}$
. As a result,
$\{\mathcal{E}_1,\mathcal{E}'_2,\mathcal{E}'_3,\ldots,\mathcal{E}'_{k+1}\}$
is a pairwise disjoint collection with
$|\mathcal{E}_1|=n-1$
,
$|\mathcal{E}'_i|=n-k-1$
. Note that we only determined the order over
$\mathcal{E}_1$
. Thus, any realization of edge scores induces a uniformly random permutation on
$\mathcal{E}_1\cup \mathcal{E'}_2\cup\cdots\cup\mathcal{E'}_{k+2}$
with the order restriction
$(1,2) \succ (1,3)\succ\cdots\succ (1,n)$
on
$\mathcal{E}_1$
.
Let
$\mathcal{E}_1$
be objects of type 0 and
$\mathcal{E'}_i$
be objects of type
$i-1$
for
$2\leq i\leq k+2$
. This satisfies all the conditions of Lemma 2. Therefore, the probability of having
$i_j$
elements of type j before the first observation of type 0 when
$k^2(\delta + 1) < i_1+\cdots+i_k < 4(k+1)^2=o(n^{1/4})$
can be approximated by the distribution
$f(\cdot)$
from (4). Hence, for sufficiently large n, we have
From what we explained earlier, the additional conditions
$i_j\geq k$
for
$1\leq j \leq k$
ensure that vertex 1 is isolated. Applying Lemma 7 then allows us to conclude that
which completes the proof.
Lemma 9. For an
$\mathit{LK}(n,k)$
graph where
$k \leq \log(n) - 3\sqrt{\log(n) \log\log(n)}$
, we have
In particular, if
$k=\lfloor t\log(n) \rfloor $
for
$t<1$
, the statement is true.
Proof. Let
$\delta = \sqrt{{6\log(k)}/{k}}$
. According to Lemma 8, for sufficiently large n and k we have
\begin{align*} & \qquad\qquad \mathbb{E}[\textrm{Number of isolated vertices}] \\[4pt] & \qquad\qquad = \mathbb{E}\Bigg[\sum_{i=1}^n I_i\Bigg] \\[4pt] & \qquad\qquad = n\mathbb{E}[I_1] \\[4pt] & \qquad\qquad = n\mathbb{P}(I_1=1) \\[4pt] & \qquad\qquad \geq \frac{n}{2}\exp((-k + 0.5)(1+\delta) + o(1)) \\[4pt] & \qquad\qquad = \exp\bigg(\log(n) - k - \sqrt{6k\log(k)} + \frac{1}{2} - \log 2 + o(1)\bigg) \\[4pt] & \qquad\qquad \geq \exp(3\sqrt{\log(n) \log\log(n)} - \sqrt{6k\log(k)}) \qquad (\mathrm{for\ sufficiently\ large}\ n) \\[4pt] & \qquad\qquad \geq \exp((3-\sqrt{6})(\sqrt{\log(n)\log\log(n)})), \end{align*}
where the right-hand side goes to infinity as
$n\rightarrow\infty$
.
To establish the desired result that there is at least one isolated vertex with high probability, it is insufficient to merely show that the mean of a non-negative random variable is unbounded, as this does not directly imply that the probability of it being zero vanishes. Therefore, we utilize the second moment method [Reference Alon and Spencer1], requiring analysis of the correlation between pairs of indicator random variables that represent vertex isolation. Let us consider two arbitrary vertices, without loss of generality vertices 1 and 2. Our goal is to study
$\mathbb{P}(I_1=1, I_2=1)=\mathbb{P}(I_1 I_2=1)=\mathbb{E}[I_1 I_2]$
. To proceed, we define the following events:
-
•
$B_{1;2}:=\{2\in\mathcal{R}^{\leq k}_1\}$
, representing the event in which vertex 2 belongs to the k-most-preferred set of vertex 1. -
•
$B_{2;1}:=\{1\in\mathcal{R}^{\leq k}_2\}$
, representing the event in which vertex 1 belongs to the k-most-preferred set of vertex 2. -
•
$B_{1\cup2}:=\{\mathcal{R}^{\leq k}_1\cap\mathcal{R}^{\leq k}_2\neq\phi\}$
, identifying the event where the k-most-preferred elements of vertices 1 and 2 have intersections. -
•
$B_{1\rightarrow2}:=\{\text{there exists}\ i\in\mathcal{R}^{\leq k}_1 \textrm{ such that } (\mathcal{R}^{\leq k}_2 \cup \{2\}) \cap \mathcal{R}^{\leq k}_i \neq \phi\}$
. This represents the event where there exists a vertex i among the k-most-preferred set of vertex 1 such that vertex 2 or one of its k-most-preferred set of vertices belong to the k-most-preferred vertices of i. -
•
$B_{2\rightarrow1}:=\{\text{there exists}\ i\in\mathcal{R}^{\leq k}_2 \textrm{ such that } (\mathcal{R}^{\leq k}_1 \cup \{1\}) \cap \mathcal{R}^{\leq k}_i \neq \phi\}$
. This represents the event where there exists a vertex i among the k-most-preferred set of vertex 2 such that vertex 1 or one of its k-most-preferred set of vertices belongs to the k-most-preferred vertices of i. -
•
$B := B_{1;2}\cup B_{2;1} \cup B_{1\cup2} \cup B_{1\rightarrow 2} \cup B_{2\rightarrow 1}$
.
For ease of presentation, we omit the value 1 from the events
$\{I_1=1\}$
,
$\{I_2=1\}$
, or
$\{I_1I_2=1\}$
. For example, we will use the shorthand
$\mathbb{P}(I_1I_2):=\mathbb{P}(I_1I_2=1)$
. Furthermore, we assume
$k<\log(n)$
for the remainder of this section. We start by bounding the event B.
Lemma 10. We have the following bound for B:
$\mathbb{P}(B) \leq O(k^3/n)$
.
Proof. First, note that
$\mathbb{P}(B_{1;2})=\mathbb{P}(B_{2;1})={k}/({n-1})$
. Consequently,
$\mathbb{P}((B_{1;2} \cup B_{2;1})^c)=1-\mathbb{P}(B_{1;2} \cup B_{2;1})=1-O(k/n)$
. Now, given the event
$(B_{1;2} \cup B_{2;1})^{\mathrm{c}}=B^{\mathrm{c}}_{1;2}\cap B^{\mathrm{c}}_{2;1}$
, the sets
$\mathcal{R}_1^{\leq k}$
and
$\mathcal{R}_2^{\leq k}$
are independent and each is uniformly drawn from the collection of all subsets of size k of
$\{3, \ldots, n\}$
. Note that
$B^{\mathrm{c}}_{1\cup\,2}$
refers to the event where
$\mathcal{R}_1^{\leq k} $
and
$\mathcal{R}_2^{\leq k}$
are disjoint. Since choosing a pair of disjoint subsets of size k from
$\{3,\ldots,n\}$
can be done in
ways, and
$\mathcal{R}_1^{\leq k}$
and
$\mathcal{R}_2^{\leq k}$
given
$B^{\mathrm{c}}_{1;2}\cap B^{\mathrm{c}}_{2;1}$
are independent and uniform, we have
\begin{align} \mathbb{P}(B^{\mathrm{c}}_{1\cup\,2}\mid B^{\mathrm{c}}_{1;2}\cap B^{\mathrm{c}}_{2;1}) & = \frac{\binom{n-2}{k}\binom{n-k-2}{k}}{\binom{n-2}{k}^2} \nonumber\\[4pt] & = \frac{\binom{n-k-2}{k}}{\binom{n-2}{k}} \nonumber\\[4pt] & = \frac{(n-k-2)(n-k-3)\cdots(n-2k-1)}{(n-2)(n-3)\cdots(n-k-1)} \nonumber\\[4pt] & \geq \bigg(1-\frac{2k-1}{n-2}\bigg)^k = 1 - O(k^2/n), \end{align}
Therefore,
$\mathbb{P}(B_{1\cup\,2}\mid B^{\mathrm{c}}_{1;2}\cap B^{\mathrm{c}}_{2;1})\leq O(k^2/n)$
. As a result,
Hence, the following bound holds:
To bound the probability of
$B_{1\rightarrow2}$
, we consider
$B_{1\rightarrow2}=\{\text{there exists} i\in\mathcal{R}^{\leq k}_1 \textrm{ such that}$
$(\mathcal{R}^{\leq k}_2 \cup \{2\}) \cap \mathcal{R}^{\leq k}_i \neq \phi\}$
given
$B_{1\cup2}^{\mathrm{c}}\cap B_{1;2}^{\mathrm{c}}\cap B_{2;1}^{\mathrm{c}}$
instead. Let
$i\in\mathcal{R}_1^{\leq k}$
be a vertex. An argument similar to (8) can be used to show that the probability of
$(\mathcal{R}^{\leq k}_2 \cup \{2\})\cap \mathcal{R}^{\leq k}_i=\phi$
is bounded below by
$(1-({k+1})/({n-1}))^k$
. As a result, using the union bound, with probability at most
there exists
$i\in\mathcal{R}_1^{\leq k}$
without this property. Hence,
Now, we bound
$\mathbb{P}(B_{1\rightarrow2})$
using this and (9):
\begin{align*} \mathbb{P}(B_{1\rightarrow2}) & = \mathbb{P}(B_{1\rightarrow2} \mid (B_{1\cup2}\cup B_{1;2}\cup B_{2;1})^{\mathrm{c}}) \mathbb{P}((B_{1\cup2}\cup B_{1;2}\cup B_{2;1})^{\mathrm{c}}) \\[4pt] & \quad + \mathbb{P}(B_{1\rightarrow2} \mid B_{1\cup2}\cup B_{1;2}\cup B_{2;1}) \mathbb{P}(B_{1\cup2}\cup B_{1;2}\cup B_{2;1}) \\[4pt] & \leq O(k^3/n) + O(k^2/n) = O(k^3/n). \end{align*}
We have established the desired bound for
$B_{1\rightarrow2}$
. The same is true for
$B_{2\rightarrow1}$
in a similar way.
Lemma 11. The following two probability bounds hold:
\begin{align*} \mathbb{P}(I_1 \mid B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}) & \leq \exp\bigg({-}k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg), \\[4pt] \mathbb{P}(I_2 \mid B_{2;1}\cup B_{1\cup2}\cup B_{2\rightarrow1}) & \leq \exp\bigg({-}k + \frac{1}{2}\sqrt{k}\log(k)+ O(\sqrt{k})\bigg). \end{align*}
Proof. We will prove the first inequality; the second one will then follow by symmetry. Let
$s=\lfloor \sqrt{k}\rfloor \leq k$
. We begin by noting that any realization of
$\mathit{LK}(n,k)$
conditional on
$B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}$
still induces a uniform distribution on the set of all permutations of
$\mathcal{E}'_i=\mathcal{E}_i\setminus \{(i,j)\colon j\in\{1,2\}\cup\mathcal{R}_1^s\}$
for
$i\neq 2$
, and in general on those of
$\mathcal{E}' = \bigcup_{i\in \mathcal{R}^{\leq s}_1\setminus\{2\}} \mathcal{E'}_i$
. This is because event
$B_{1;2}$
does not affect the orders within edges in
$\mathcal{E}'$
under the
$i\neq 2$
assumption. Moreover, since
$(i,2)\notin \mathcal{E}_i'$
,
$B_{1\cup\,2}$
and
$B_{1\rightarrow2}$
only impose order restrictions on edges of the form
$\{(j,2)\colon j\neq 2\} $
, which are disjoint from
$\mathcal{E}'$
. Hence, considering permutations of the union
$\mathcal{E}_1\cup\mathcal{E}'$
given
$B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}$
, there are only order restrictions on the elements of
$\mathcal{E}_1$
. Note that since
$\mathcal{E}_1,\mathcal{E}'_{i_1},\mathcal{E}'_{i_2},\ldots$
are disjoint for
$i_j\in \mathcal{R}^{\leq s}_1\setminus\{2\}$
, elements of
$\mathcal{E}_1,\mathcal{E}'_{i_1},\mathcal{E}'_{i_2},\ldots$
can be assigned types
$0,1,2,\ldots,s'$
, respectively, where
$s'=s-1$
or s depending on whether
$2\in\mathcal{R}_1^{\leq s}$
or not.
We now discuss a necessary condition on these permutations for which vertex 1 is isolated. Suppose that vertex 1 is isolated conditional on
$B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}$
. Therefore, for any
$i\in\mathcal{R}_1^{\leq k}$
, and specifically for any
$i\in\mathcal{R}_1^{\leq s}\setminus\{2\}$
, we have
$(1,i)\notin \mathcal{R}_i^{\leq k}$
. Equivalently, there exist k elements of
$\mathcal{E}_i\setminus\{(1,i)\}$
appearing before (1, i) in the permutation. Thus, at least
$k-s-2$
elements of
$\mathcal{E}'_i$
appear earlier than (1, i). This holds for all
$i\in\mathcal{R}_1^{\leq s}\setminus\{2\}$
, and since
$\mathcal{E}^{\leq s}_1=\{(1,i)\colon i\in\mathcal{R}_1^{\leq s}\}$
are the first s elements of type 0 placed in this permutation, we conclude that there are at most s elements of type 0 in the first
$(s-1)(k-s-2)$
elements of this permutation. Lemma 3 bounds the probability of the event above by
which completes the proof.
Lemma 12.
$\mathbb{P}(I_1I_2\cap B)\leq \exp\big({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\big)$
.
Proof. We establish the bound through Lemmas 10 and 11.
\begin{align*} \mathbb{P}(I_1I_2\cap B) & \leq \mathbb{P}(I_1I_2 \cap (B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2})) + \mathbb{P}(I_1I_2 \cap (B_{2;1}\cup B_{1\cup2}\cup B_{2\rightarrow1})) \\[6pt] & \leq \mathbb{P}(I_1 \cap (B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2})) + \mathbb{P}(I_2 \cap (B_{2;1}\cup B_{1\cup2}\cup B_{2\rightarrow1})) \\[6pt] & = \mathbb{P}(I_1 \mid B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}) \mathbb{P}(B_{1;2}\cup B_{1\cup2}\cup B_{1\rightarrow2}) \\[6pt] & \quad + \mathbb{P}(I_2 \mid B_{2;1}\cup B_{1\cup2}\cup B_{2\rightarrow1}) \mathbb{P}(B_{2;1}\cup B_{1\cup2}\cup B_{2\rightarrow 1}) \end{align*}
\begin{align*} & \leq O\bigg(\frac{k^3}{n}\bigg)\exp\bigg({-}k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg) \qquad (\text{Lemmas }{10}\text{ and }{11}) \\[4pt] & = \exp\bigg({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log (k) + O(\sqrt{k})\bigg). \end{align*}
Thus, the result holds.
Lemma 13. We have the conditional independence
$\mathbb{P}(I_1 I_2 \mid B^{\mathrm{c}}) = \mathbb{P}(I_1 \mid B^{\mathrm{c}})\mathbb{P}(I_2 \mid B^{\mathrm{c}})$
.
Proof. Note that
$I_1$
holds whenever, for each
$1\leq i\leq k$
, there are at least k elements
$j\in[n]\setminus\{1, R_1^i\}$
such that
$V(R_1^i, j) > V(R_1^i, 1)$
. Assuming
$B^{\mathrm{c}}$
, this is equivalent to there being at least k elements
$j_1\in [n] \setminus \{1,2\} \setminus \{R_2^1,\ldots,R_2^k\}$
with this property. Repeating this argument for
$I_2$
, for each
$1\leq i\leq k$
, there must be at least k elements
$j_2\in [n] \setminus \{1,2\} \setminus \{R_1^1, R_1^2,\ldots, R_1^k\}$
with
$V(R_2^i, j_2) > V(R_2^i, 2)$
. Ruling out the possibility of choosing
$j_2,j_1$
from
$\{1,2,R_1^1, R_1^2,\ldots, R_1^k\}$
and
$\{1,2,R_2^1,\ldots,R_2^k\}$
respectively implies that there is no common edge among
$(R_1^i,1)$
,
$(R_1^i,j_1)$
,
$(R_2^i,2)$
, and
$(R_2^i, j_2)$
. Since the scores are chosen independently, and the inequality conditions for the isolation of vertex 1 and that of vertex 2 are disjoint, we obtain conditional independence.
Lemma 14. Let
$t'<-0.5$
be a real number. Assume k grows to infinity as
$n\rightarrow\infty$
with
$k\leq \log(n) + t' \log\log(n)\sqrt{\log(n)}$
. Then
Proof. We begin with the equality
It follows that
\begin{align*} \frac{\mathbb{P}(I_1 \cap B)}{\mathbb{P}(I_1)\mathbb{P}(B^{\mathrm{c}})} & \leq O\bigg(\frac{k^3}{n}\bigg){\mathrm{e}}^{k(1+\sqrt{{6\log(k)}/{k}})+o(1)} \\[5pt] & = \exp(k+\sqrt{6k\log(k)} + 3\log(k) - \log(n) + O(1)) \\[5pt] & \leq \exp(t'\log\log(n)\sqrt{\log(n)} + \sqrt{6k\log(k)} + 3\log(k) + O(1)) \\[5pt] & \leq \exp(t'\log(k)\sqrt{k} + \sqrt{6k\log(k)} + 3\log(k) + O(1)) \overset{{k\rightarrow\infty}}{\rightarrow} 0, \end{align*}
where we use the fact that
$k \leq \log(n)$
. Considering the fact that
${1}/{\mathbb{P}(B^{\mathrm{c}})}\rightarrow 1$
, (10) implies that
It follows from Lemmas 12 and 13 that
\begin{align*} \mathbb{P}(I_1I_2) & = \mathbb{P}(I_1I_2 \cap B^{\mathrm{c}}) + \mathbb{P}(I_1I_2 \cap B) \\[4pt] & \leq \mathbb{P}(I_1I_2 \cap B^{\mathrm{c}}) + \exp\bigg({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg) \\[4pt] & \leq \mathbb{P}(I_1I_2 \mid B^{\mathrm{c}}) + \exp\bigg({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg) \\[4pt] & = \mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2 \mid B^{\mathrm{c}}) + \exp\bigg({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg). \end{align*}
Dividing both sides of this by
$\mathbb{P}(I_1)\mathbb{P}(I_2)=\mathbb{P}(I_1)^2$
and using Lemma 8, we have
\begin{align*} & \frac{\mathbb{P}(I_1I_2)}{\mathbb{P}(I_1)\mathbb{P}(I_2)} \\[6pt] & \leq \frac{\mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2\mid B^{\mathrm{c}})} {\mathbb{P}(I_1)\mathbb{P}(I_2)} + \frac{\exp\big({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\big)} {{\mathbb{P}(I_1)\mathbb{P}(I_2)}} \\[8pt] & \leq \frac{\mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2\mid B^{\mathrm{c}})} {\mathbb{P}(I_1)\mathbb{P}(I_2)} + \frac{\exp\big({-}\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\big)} {\exp\big({-}2k(1+\sqrt{{6\log(k)}/{k}}) + o(1)\big)} \\[8pt] & \leq \frac{\mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2\mid B^{\mathrm{c}})} {\mathbb{P}(I_1)\mathbb{P}(I_2)} + \exp\bigg({-}\log(n) + k + \frac{1}{2}\sqrt{k}\log(k) + \sqrt{24k\log(k)} + O(\sqrt{k})\bigg) \\[8pt] & \leq \frac{\mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2\mid B^{\mathrm{c}})} {\mathbb{P}(I_1)\mathbb{P}(I_2)} + \exp\bigg(t'\log\log(n)\sqrt{\log(n)}+\frac{1}{2}\sqrt{k}\log(k)+\sqrt{24k\log(k)}+O(\sqrt{k})\bigg) \\[8pt] & \leq \frac{\mathbb{P}(I_1\mid B^{\mathrm{c}})\mathbb{P}(I_2\mid B^{\mathrm{c}})} {\mathbb{P}(I_1)\mathbb{P}(I_2)} + \exp((0.5 + t')\log(k)\sqrt{k} + \sqrt{24k\log(k)} + O(\sqrt{k})). \end{align*}
Equation (11) guarantees that the first term on the right-hand side converges to 1. The second term vanishes as
$k\rightarrow\infty$
as a result of
$t'<-0.5$
. Since
$\mathbb{E}[I_1]=\mathbb{P}(I_1)$
and
$\mathbb{E}[I_1I_2]=\mathbb{P}(I_1I_2)$
, this completes the proof.
Theorem 3. The random graph model
$\mathit{LK}(n,k)$
with
$k \leq \log(n) + t'\log\log(n)\sqrt{\log(n)}$
for
$t'<-\tfrac{1}{2}$
is not connected with high probability. In particular, if
$k=\lfloor t\log(n)\rfloor$
for
$0<t<1$
, the graph is disconnected with high probability.
Proof. Using both Lemmas 9 and 14, we can apply the second moment method. We begin with
\begin{equation*} \mathbb{E}\Bigg[\Bigg(\sum_{i=1}^n I_i\Bigg)^2\Bigg] - \mathbb{E}\Bigg[\sum_{i=1}^n I_i\Bigg]^2 = \mathrm{Var}\Bigg[\sum_{i=1}^n I_i\Bigg] \geq \Bigg(0 - \mathbb{E}\Bigg[\sum_{i=1}^n I_i\Bigg]\Bigg)^2\mathbb{P}\Bigg(\sum_{i=1}^n I_i = 0\Bigg). \end{equation*}
Therefore,
\begin{align*} \mathbb{P}\Bigg(\sum_{i=1}^n I_i = 0\Bigg) & \leq \frac{\mathbb{E}\big[\big(\sum_{i=1}^n I_i\big)^2\big] - \mathbb{E}\big[\sum_{i=1}^n I_i\big]^2}{\mathbb{E}\big[\sum_{i=1}^n I_i\big]^2} \\[4pt] & \leq \frac{\mathbb{E}\big[\big(\sum_{i=1}^n I_i \big)^2\big]}{\mathbb{E}\big[\sum_{i=1}^n I_i\big]^2} - 1 \\[4pt] & = \frac{\sum_{i=1}^n\mathbb{E}[I_i^2]+\sum_{i,j,i\neq j}\mathbb{E}[I_iI_j]}{n^2\mathbb{E}[I_1]^2} - 1 \\[4pt] & = \frac{n\mathbb{E}[I_1] + n(n-1)\mathbb{E}[I_1I_2]}{n^2\mathbb{E}[I_1]^2} - 1 = \frac{1}{n\mathbb{E}[I_1]} + \frac{(n-1)\mathbb{E}[I_1I_2]}{n\mathbb{E}[I_1]^2} - 1. \end{align*}
Now it follows from Lemma 9 that the first fraction vanishes as
$n\rightarrow\infty$
. That is, for any
$\varepsilon>0$
, there exists
$N_1>0$
such that
${1}/{n\mathbb{E}[I_1]} \leq {\varepsilon}/{2}$
for all
$n>N_1$
. Moreover, Lemma 14 implies that the second fraction will be sufficiently close to 1. Indeed, there exists
$N_2>0$
such that, for all
$n>N_2$
, we have
${\mathbb{E}[I_1I_2]}/{\mathbb{E}[I_1]^2} \leq 1 + {\varepsilon}/{2}$
. Hence, for any given
$\varepsilon>0$
, we find
$N=\max(N_1,N_2)$
such that, for any
$n>N$
,
\begin{equation*} 0 \leq \mathbb{P}\Bigg(\sum_{i=1}^n I_i = 0\Bigg) \leq \frac{1}{n\mathbb{E}[I_1]} + \frac{(n-1)\mathbb{E}[I_1I_2]}{n\mathbb{E}[I_1]^2} - 1 \leq \frac{\varepsilon}{2} + 1 + \frac{(n-1)\varepsilon}{2n} - 1 \leq \varepsilon. \end{equation*}
This shows that
$\mathbb{P}\big(\sum_{i=1}^n I_i = 0\big)\rightarrow0$
as
$n\rightarrow\infty$
. Therefore, the probability of having no isolated vertex converges to 0, ensuring the existence of an isolated vertex with high probability as n grows to infinity.
5. Connectivity of
$\mathit{LK}(\mathit{n},\mathit{k})$
graphs for
$\mathit{t}>\textbf{1}$
A common method to prove connectivity of random graphs is by locating (with high probability) an Erdős–Rényi subgraph consisting of all the vertices. Then, the celebrated Erdős–Rényi result [Reference Erdős and Rényi6, Reference Erdős and Rényi7, Reference Rényi13] guarantees connectivity if the probability of an edge being present is
$p = {t\log(n)}/{n}$
, where
$t>1$
. This method, used in [Reference La and Kabkab10], proves the connectivity of
$\mathit{LK}(n,k)$
for
$k=t\log(n)$
where
$t>C=2.4625$
. The link between Erdős–Rényi and
$\mathit{LK}(n,k)$
is made through a concentration property of order statistics. The authors of [Reference La and Kabkab10] assume that the edges are independently scored by
$\mathrm{Exp}(1)$
random variables, which results in the presence of edges whose scores are greater than or equal to
$\log((n-1)/(t\log(n))) + \sqrt{2/t}$
with high probability. Therefore, if
$L_{i,j}$
is a random variable distributed as
$\mathrm{Exp}(1)$
, then
It can be easily shown that
$t{\mathrm{e}}^{-\sqrt{{2}/{t}}}>1$
for
$t>2.4625$
. This provides an independent possibility of connection for all edges with the probability of
$t'\log (n)/(n-1)$
for a
$t'=t{\mathrm{e}}^{-\sqrt{{2}/{t}}}>1$
. Thereafter, the connectivity result for Erdős–Rényi random graphs [Reference Erdős and Rényi6, Reference Erdős and Rényi7, Reference Rényi13] implies the connectivity of
$\mathit{LK}(n,k)$
.
We prove the connectivity of
$\mathit{LK}(n,k)$
for all
$t>1$
in two steps. First, we rule out the possibility of having components of size O(1). Second, we apply the idea from [Reference La and Kabkab10] described above to find an Erdős–Rényi graph with all edges contained in
$\mathit{LK}(n,k)$
. As opposed to [Reference La and Kabkab10], we do not restrict t to find a
$t'>1$
. Hence, the resulting Erdős–Rényi graph is not necessarily connected. However, by modifying the original proof of the connectivity of Erdős–Rényi graphs for
$t'>1$
from [Reference Rényi13], we can prove a weaker result for an arbitrary
$t'>1$
that helps us deduce the connectivity result.
Theorem 4. Let
$\kappa\geq 0$
be a non-negative integer. Consider random graphs of class
$\mathit{LK}(n,k)$
where
$k \geq \log(n) + t'\log\log(n)\sqrt{\log(n)}$
for a real number
$t'>\tfrac{1}{2}$
. Then
Proof. We will bound the probability that vertex 1 has degree less than
$\kappa$
with the same technique used in the proof of Lemma 11. Let
$s=\lfloor \sqrt{k}\rfloor \leq k$
and, for
$i\in \mathcal{R}^{\leq s}_1$
,
$\mathcal{E'}_{i} = \mathcal{E}_{i} \setminus \{(i,j)\colon j\in\mathcal{R}^{\leq s}_1\cup\{1\}\}$
. Now any realization of
$\mathit{LK}(n,k)$
induces a uniform distribution on the set of all permutations of the union
$\mathcal{E}' = \mathcal{E}_1 \bigcup_{i\in \mathcal{R}^{\leq s}_1} \mathcal{E'}_i$
, with some order restrictions only within the elements of
$\mathcal{E}_1$
. This union is a partition in which
$|\mathcal{E}_1|=n-1$
, and
$|\mathcal{E'}_i|=n-s-1$
for all
$i\in \mathcal{R}^{\leq s}_1$
. We assign type 0 to the elements of
$\mathcal{E}_1$
and type i to the elements of
$\mathcal{E'}_i$
for
$i\in\mathcal{R}_1^{\leq s}$
. Therefore, any realization of
$\mathit{LK}(n,k)$
induces a uniformly random permutation over
$\mathcal{E}'$
with some order restrictions on the elements of type 0.
A necessary condition to ensure that the degree of vertex 1 is bounded above by
$\kappa$
is that for at least
$s-\kappa$
elements and
$i\in \mathcal{R}^{\leq s}_1$
, at least k edges of
$\mathcal{E}_{i}$
appear before (1, i) in the permutation. As a result, those k edges appear before the sth element of
$\mathcal{E}_{1}$
too. This implies there must be at least
$k-s-1$
edges of
$\mathcal{E'}_{i}$
before the sth element of
$\mathcal{E}_{1}$
. As a result, in the first
$(s-\kappa)(k-s-1)$
elements of this permutation there are at most
$s-1$
elements of type 0. Lemma 3 bounds the probability of the event above by
$\exp\big({-}k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\big)$
. Then, using the union bound for sufficiently large n, we have
\begin{align*} \mathbb{P}(\text{there exists}\ v\colon\deg(v) \leq \kappa) & \leq n \times \mathbb{P}(\text{vertex 1 is of degree at most}\ \kappa) \\ & \leq n\exp\bigg({-}k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg) \\ & = \exp\bigg(\log(n) - k + \frac{1}{2}\sqrt{k}\log(k) + O(\sqrt{k})\bigg). \end{align*}
If
$k>O(\log(n))$
, then the right-hand side obviously vanishes. In the case
$k=O(\log(n))$
, we use the fact that
$k - \frac{1}{2}\sqrt{k}\log(k)+O(\sqrt{k})$
is increasing for sufficiently large k, and replace k by
$\log(n)+t'\log\log(n)\sqrt{\log(n)}$
in the last equality to obtain
Hence, the convergence holds.
Theorem 4 excludes the possibility of having components of size O(1). In other words, the following corollary holds.
Corollary 1. Consider the random graphs model of class
$\mathit{LK}(n,k)$
with
for a real number
$t'>\tfrac{1}{2}$
. Then
$\mathbb{P}(\textit{there exists a component of size at most}\ \kappa) \rightarrow 0$
for any fixed
$\kappa$
as
$n\rightarrow\infty$
. In particular, when
$k=\lfloor t\log(n)\rfloor$
for
$t>1$
, the limit above is still true.
Proof. In the case of a component of size
$\kappa$
there must be a vertex of degree at most
$\kappa$
, which is impossible with high probability due to Theorem 4.
Next, we establish a proof based on finding Erdős–Rényi sub- and super-graphs of
$\mathit{LK}(n,k)$
with both containing all the vertices. Recall that the order statistics of a set of i.i.d. random variables
$\{X_1,X_2,\ldots,X_n\}$
is defined as their non-increasing rearrangement
$X^{(1)}\geq X^{(2)}\geq \cdots \geq X^{(n)}$
. Recall that the kth order statistic of
$\mathcal{V}_i$
is denoted by
$V(i,R_i^k)$
.
Theorem 5. For the random graph model
$\mathit{LK}(n,k)$
with
$k=\lfloor t \log(n)\rfloor$
, there is no component of size
$\lceil{8.24}/{t}\rceil \leq r \leq \lfloor{n}/{2}\rfloor$
with high probability.
Proof. First, we refer to [Reference La and Kabkab10, Lemma A.1], which states that if the scores V(i, j) are independently distributed as exponential random variables with parameter 1, and we define
then
$\mathbb{P}(A_n)\rightarrow 1$
as
$n\rightarrow\infty$
. We let
Next, note that
\begin{align*} \bar{p} & := \mathbb{P}(V(i,j) > \bar{l}) = \frac{t\log(n)}{n-1}{\mathrm{e}}^{-\sqrt{2}} \approx \frac{0.2431t\log(n)}{n-1}, \\[3pt] \underline{p} & := \mathbb{P}(V(i,j) > \underline{l}) = \frac{t\log(n)}{n-1}{\mathrm{e}}^{\sqrt{2}} \approx \frac{4.1133t\log(n)}{n-1}. \end{align*}
Additionally, if the graph meets condition
$A_n$
(which holds with high probability), then
$V(i,j) > \bar{l}$
indicates that the edge (i, j) exists, while
$V(i,j) < \underline{l}$
ensures that the edge (i, j) does not exist.
Let the graph satisfy condition
$A_n$
. In order to have a component of size r, we must choose r vertices containing at least one spanning tree. Moreover, according to Cayley’s formula, there are
$r^{r-2}$
possible trees with r vertices. Thus, the
$r-1$
edges of the tree must have scores no less than
$\underline{l}$
, which occurs with probability
$\underline{p}^{r-1}$
due to independence. Additionally, we require that those r vertices are not connected to the rest of the graph, implying that the scores for the intermediate edges between the component and the rest of the graph must be less than
$\bar{l}$
. Note that both of these are necessary conditions. Counting the number of possible components yields the following upper bound:
\begin{align*} \Pi & = \mathbb{P}\big(\text{there exists a component of size between}\ \lceil8.24/t\rceil\ \text{and}\ \lfloor n/2\rfloor\big) \\[3pt] & \leq \sum_{r=\lceil8.24/t\rceil}^{\lfloor n/2\rfloor}\binom{n}{r}r^{r-2}\underline{p}^{r-1}(1-\bar{p})^{r(n-r)}. \end{align*}
Substituting
$\binom{n}{r} < {n^r}/{r!} < n^r/\sqrt{2\pi r}(r/{\mathrm{e}})^r$
implies that
\begin{align*} \Pi & < \frac{1}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\lfloor n/2\rfloor} n^rr^{-r-1/2}{\mathrm{e}}^rr^{r-2}\underline{p}^{r-1}(1-\bar{p})^{r(n-r)} \\[3pt] & < \frac{1}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\lfloor n/2\rfloor} {\mathrm{e}}^r\frac{n^r}{\underline{p}}r^{-5/2}\underline{p}^{r}{\mathrm{e}}^{-r(n-r)\bar{p}} \qquad (\text{using}\ 1-x < {\mathrm{e}}^{-x}) \end{align*}
\begin{align*} & < \frac{n}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\lfloor n/2\rfloor} n^r{\mathrm{e}}^r\underline{p}^{r}{\mathrm{e}}^{-r\bar{p}n/2} \qquad (\text{using}\ 1/\underline{p} < n,\ n-r>n/2,\ r^{-5/2} < 1) \\[10pt] & < \frac{n}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\lfloor n/2\rfloor}{\mathrm{e}}^{r(1 + \log(n\underline{p}) - n\bar{p}/2)}. \end{align*}
As t is a fixed number, the dominant term in
$1 + \log(n\underline{p}) - n\bar{p}/2$
is
$- n\bar{p}/2\approx -0.1216 t\log(n)$
. As a result, for sufficiently large n,
$1 + \log(n\underline{p}) - n\bar{p}/2 < -0.1215t\log(n)$
. Thus, we have
\begin{align*} \Pi < \frac{n}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\infty}{\mathrm{e}}^{-0.1215tr\log(n)} & = \frac{n}{\sqrt{2\pi}\,}\sum_{r=\lceil{8.24}/{t}\rceil}^{\infty}n^{-0.1215tr} \\[12pt] & = \frac{n^{1 - 0.1215t\lceil{8.24}/{t}\rceil}}{\sqrt{2\pi}(1-n^{-0.1215t})} \\[12pt] & \leq \frac{n^{-0.00116}}{\sqrt{2\pi}(1-n^{-0.1215t})} \overset{{n\rightarrow\infty}}{\rightarrow} 0. \end{align*}
Therefore, with high probability there is no component of size between
$\lceil 8.24/t\rceil$
and
$n/2$
.
A key feature of Theorem 5 is its validity for any positive t. This means that if
$k\approx t\log(n)=\Theta(\log(n))$
, we should only examine components of size less than
$\lceil{8.24}/{t}\rceil=O(1)$
to study the disconnectedness of the graph. Since Corollary 1 addresses components of size O(1), we now assert the main theorem.
Theorem 6. The random graph model
$\mathit{LK}(n,k)$
with
$k \geq \log(n) + t'\log\log(n)\sqrt{\log(n)}$
for a real number
$t'>\tfrac{1}{2}$
is connected with high probability. In particular, if
$k=\lfloor t\log(n)\rfloor$
for
$t>1$
, connectivity holds with high probability.
Proof. If a graph from class
$\mathit{LK}(n,k)$
is disconnected, then it should have at least a component of size s where
$1 \leq s \leq \lfloor n/2\rfloor$
. Since
$\lceil{8.24}/{1}\rceil=9$
, Theorem 5 rules out the possibility of having components of size between 9 and
$\lfloor n/2\rfloor$
with high probability. Moreover, Corollary 1 guarantees that the probability of having components of size at most 8 vanishes when n grows to infinity. Therefore, the graph will be connected with high probability.
6. Average degree of
$\mathit{LK}(\mathit{n},\mathit{k})$
graphs
Here, we discuss another set of results for
$\mathit{LK}(n,k)$
random graphs. From the construction it is immediate that the degree sequence is bounded above by k. Therefore, the average degree is also constrained by k. However, we will show that for a large range of k (as a function of n), this number is very close to k by specifying the error term. Finally, we compare our result with the results on the sparse case in [Reference Moharrami, Subramanian, Liu and Sundaresan11].
Before presenting the main theorem we recall the negative binomial distribution. Imagine an urn with an infinite number of red and blue balls, where each draw results in a red ball with probability p. The probability of drawing j blue balls before the kth red ball is a negative binomial distribution given by
$\mathbb{P}(X_k=j)=\binom{k+j-1}{j}p^k(1-p)^{j}$
. We need to introduce two combinatorial lemmas before stating the main theorem.
Lemma 15.
\begin{align*}\sum_{j=0}^{k-1} \binom{k+j-1}{j} \frac{1}{2^{k+j}}=\frac{1}{2}.\end{align*}
Proof. Let
$X_k$
denote the negative binomial random variable with
$p=\frac{1}{2}$
. Note that
$\mathbb{P}(X_k < k)=\sum_{j=0}^{k-1}\binom{k+j-1}{j}({1}/{2^{k+j}})$
is the probability of observing the kth red ball earlier than the kth blue ball in the infinite urn model. By symmetry this value must be
$\frac12$
.
Lemma 16.
\begin{align*} \sum_{j=0}^{k-1}\binom{k+j-1}{j}\bigg(\frac{j}{2^{k+j}}\bigg) = \frac{k}{2}-\frac{(2k-1)}{2^{2k-1}}\binom{2k-2}{k-1}. \end{align*}
Proof. Using Lemma 15, we can write the sum above as
\begin{align*} A & = \sum_{j=0}^{k-1}\binom{k+j-1}{j}\bigg(\frac{j}{2^{k+j}}\bigg) \\[4pt] & = \sum_{j=1}^{k-1}\binom{k+j-2}{j-1}\bigg(\frac{k+j-1}{2^{k+j}}\bigg) \\[4pt] & = \bigg(\frac{k}{2}\bigg)\sum_{j=1}^{k-1}\binom{k+j-2}{j-1}\frac{1}{2^{j+k-1}} + \bigg(\frac{1}{2}\bigg)\sum_{j=1}^{k-1}(j-1)\binom{k+j-2}{j-1}\frac{1}{2^{j+k-1}} \\[4pt] & = \bigg(\frac{k}{2}\bigg)\bigg[\frac{1}{2}-\binom{2k-2}{k-1}\frac{1}{2^{2k-1}}\bigg] + \bigg(\frac{1}{2}\bigg)\bigg[A - (k-1)\binom{2k-2}{k-1}\frac{1}{2^{2k-1}}\bigg] \\[4pt] & = \frac{k}{4} + \frac{A}{2} - \frac{(2k-1)}{2^{2k-2}}\binom{2k-2}{k-1}. \end{align*}
We can solve the resulting equation for A to get the desired expression.
Theorem 7. In the model
$\mathit{LK}(n,k)$
, suppose D represents the degree of a randomly chosen vertex. If
$k=o(\sqrt{n})$
then we have the asymptotic
Proof. Without loss of generality, assume that the randomly chosen vertex is vertex 1 and
$R_1^i = i+1$
for all
$1\leq i\leq n-1$
. Let
$E_i$
denote the indicator function for whether vertex 1 is connected to vertex
$i+1$
. To prevent the graph from having edge
$(1,i+1)$
, vertex
$i+1$
must disagree on vertex 1. This is equivalent to at least k elements of
$\mathcal{E}_{i+1}$
appearing before the ith element of
$\mathcal{E}_1$
, which is
$(1,i+1)$
, in the permutation. Hence, there could be at most i elements of
$\mathcal{E}_1$
before the kth element of
$\mathcal{E}_{i+1}$
. Any realization of
$\mathit{LK}(n,k)$
induces a uniformly random permutation on
$\mathcal{E}_1\cup \mathcal{E}_{i+1}$
subject to the order restriction
$(1,2)\succ(1,3)\succ\cdots\succ(1,n)$
. By employing a similar approach to the proof of Lemma 2, we can use combinatorial methods to calculate the probability of observing j elements of
$\mathcal{E}_1$
earlier than the kth element of
$\mathcal{E}_{i+1}$
for
$j\leq k$
as follows:
\begin{align*} \binom{k+j-1}{j}\frac{(n-1)(n-2)\cdots(n-k-j)}{(2n-3)(2n-4)\cdots(2n-k-j-2)} & = \binom{k+j-1}{j}\bigg(\frac{n+O(k)}{2n+O(k)}\bigg)^{k+j} \\ & = \binom{k+j-1}{j}\bigg(\frac{1}{2}\bigg)^{k+j}(1 + O(k^2/n)). \end{align*}
Varying
$0\leq j \leq i-1$
, it follows that, for each
$1\leq i \leq k$
,
\begin{align*}\mathbb{P}(E_i = 0) = \Bigg[\sum_{j=0}^{i-1}\binom{k+j-1}{j}\frac{1}{2^{k+j}}\Bigg](1+O(k^2/n)).\end{align*}
We now compute
$\mathbb{E}[D]$
using Lemmas 15 and 16:
\begin{align*} \mathbb{E}[D] = \sum_{i=1}^{n-1}\mathbb{E}[E_i] = \sum_{i=1}^{n-1}\mathbb{P}(E_i = 1) & = \sum_{i=1}^{k}\mathbb{P}(E_i = 1) \\ & = k - \sum_{i=1}^{k}\mathbb{P}(E_i = 0) \\ & = k - \sum_{i=1}^{k}\Bigg[\sum_{j=0}^{i-1}\binom{k+j-1}{j}\frac{1}{2^{k+j}}\Bigg](O(k^2/n)+1) \\ & = k - \Bigg[\sum_{j=0}^{k-1}\binom{k+j-1}{j}\frac{k-j}{2^{k+j}}\Bigg](O(k^2/n)+1) \\ & = k - \Bigg[\frac{k}{2} - \sum_{j=0}^{k-1}\binom{k+j-1}{j}\frac{j}{2^{k+j}}\Bigg](O(k^2/n)+1) \\ & = k - \bigg[\frac{(2k-1)}{2^{2k-1}}\binom{2k-2}{k-1}\bigg](O(k^2/n)+1). \end{align*}
This completes the proof.
Remark 2. As mentioned earlier, [Reference Moharrami, Subramanian, Liu and Sundaresan11] considers the preference threshold to be a random variable independently chosen per vertex using a distribution P over
$\mathbb{N}$
with finite mean instead of just a fixed k for the potential number of neighbors of a randomly chosen vertex. Further, [Reference Moharrami, Subramanian, Liu and Sundaresan11, Theorem 5.1] specifies the following formula for the average degree in the limit of n going to infinity:
\begin{equation} \mathbb{E}[D] = \sum_{i=1}^{\infty}\sum_{j=1}^{\infty}P(i)P(j)\int_0^{\infty}\bar{F}_i(x)\bar{F}_j(x)\,{\mathrm{d}} x, \end{equation}
where
$\bar{F}_i$
denotes the complementary cumulative distribution function of
$\textrm{Erlang}(\cdot\, ;i,1)$
(the Erlang distribution of shape i and rate 1).
In the case of bilateral preference graphs with preference threshold parameter a fixed k, the probability distribution P becomes a delta mass function on k, i.e.
$P(i)=1$
if
$i=k$
, and
$P(i)=0$
otherwise. In addition, the mean being finite as n goes to infinity implies that k must be finite and cannot grow to infinity. The sum in (13) now simplifies to
$\mathbb{E}[D] = \int_0^{\infty}\bar{F}_k^2(x)\,{\mathrm{d}} x$
. Then, using
$\bar{F}_k(x)=\sum_{i=0}^{k-1}{\mathrm{e}}^{-x}x^i/{i!}$
, distributing the square, and exchanging the finite sum with the integral, we obtain
\begin{align*} \mathbb{E}[D] & = \sum_{0\leq i,j\leq k-1}\int_0^{\infty}\frac{x^{i+j}}{i!j!}{\mathrm{e}}^{-2x}\,{\mathrm{d}} x \\[4pt] & = \sum_{0\leq i,j\leq k-1}\int_0^{\infty}\frac{(2x)^{i+j}}{2^{i+j}i!j!}{\mathrm{e}}^{-2x}\,{\mathrm{d}} x \\[4pt] & = \sum_{0\leq i,j\leq k-1}\frac{\Gamma(i+j+1)}{2^{i+j+1}i!j!} = \frac{1}{2}\sum_{0\leq i,j\leq k-1} \frac{1}{2^{i+j}} \binom{i+j}{i}.\end{align*}
We substitute
$s=i+j$
and denote the negative binomial random variable with parameter
$p=\frac12$
to represent the probability of the kth red ball appearing after observing
$X_k$
blue balls. This leads us to
\begin{align*} \mathbb{E}[D] & = \frac{1}{2}\Bigg[\sum_{s=0}^{k-1}\sum_{i=0}^s\frac{1}{2^{s}}\binom{s}{i} + \sum_{s=k}^{2k-2}\sum_{i=s-k+1}^{k-1}\frac{1}{2^s}\binom{s}{i}\Bigg] \\[4pt] & = \frac{1}{2}\Bigg[k + \sum_{s=0}^{k-2}\Bigg(1-2\sum_{i=0}^{s}\frac{1}{2^{s+k}}\binom{s+k}{i}\Bigg)\Bigg] \\[4pt] & = \frac{1}{2}\Bigg[k + \sum_{s=0}^{k-2}(1-2\mathbb{P}(X_k\leq s ))\Bigg] \\[4pt] & = \frac{1}{2}\Bigg[k + \sum_{s=0}^{k-2}(2\mathbb{P}(X_k> s )-1)\Bigg] \\[4pt] & = \frac{1}{2}[1 + 2\mathbb{E}[X_k\mathbf{1}_{X_k\leq k-1}] + 2(k-1)\mathbb{P}(X_k\geq k)] \\[4pt] & = \frac{1}{2}\Bigg[1 + \sum_{j=0}^{k-1}\binom{k+j-1}{j}\frac{2j}{2^{k+j}} + 2(k-1)(1-\mathbb{P}(X_k < k))\Bigg] \\[4pt] & = \frac{k}{2} + \sum_{j=0}^{k-1}\binom{k+j-1}{j}\frac{j}{2^{k+j}} \qquad (\text{using Lemma }{15}) \\[4pt] & = k - \bigg[\frac{(2k-1)}{2^{2k-1}}\binom{2k-2}{k-1}\bigg] \qquad (\text{using Lemma }{16}).\end{align*}
Therefore, we reproduce the same formula for the average degree as in Theorem 7 using the result in [Reference Moharrami, Subramanian, Liu and Sundaresan11]. We discussed this remark with the assumption of sparseness. However, Theorem 7 only needs
$k=o(\sqrt{n})$
, which covers both the sparse and non-sparse regimes.
Next, we determine the asymptotic behavior of the mean degree as both n and k go to infinity with a constraint on how fast k grows relative to n.
Theorem 8. In the model
$\mathit{LK}(n,k)$
, assume that
$k=o(n^{1/3})$
grows to infinity as n goes to infinity. Suppose D represents the degree of a randomly chosen vertex. We have the following asymptotic behavior of the average degree:
Proof. We simply use Stirling’s approximation
to find an asymptotic expression for
$\binom{2k-2}{k-1}$
:
\begin{align*} \binom{2k-2}{k-1} = \frac{(2k-2)!}{(k-1)!^2} & = \frac{\sqrt{2\pi(2k-2)}(({2k-2})/{{\mathrm{e}}})^{2k-2}(1+{1}/({12(2k-2)})+o({1}/{k}))} {2\pi(k-1)(({k-1})/{{\mathrm{e}}})^{2k-2}(1+{1}/({12(k-1)})+o({1}/{k}))^2} \\[5pt] & = \frac{\sqrt{2\pi(2k-2)}2^{2k-2}(1+{1}/({12(2k-2)})+o({1}/{k}))} {2\pi(k-1)(1+{1}/({12(k-1)})+o({1}/{k}))^2} \\[5pt] & = \frac{2^{2k-2}(1+{1}/({24(k-1)})+o({1}/{k}))}{\sqrt{\pi(k-1)}(1+{1}/({6(k-1)})+o({1}/{k}))} \\[5pt] & = \bigg(\frac{2^{2k-2}}{\sqrt{\pi(k-1)}}\bigg)\bigg(1 - \frac{1}{8(k-1)} + o\bigg(\frac{1}{k}\bigg)\bigg). \end{align*}
Substituting this into (12) implies that
\begin{align*} \mathbb{E}[D] & = k - \bigg[\frac{(2k-1)}{2^{2k-1}}\binom{2k-2}{k-1}\bigg](O(k^2/n)+1) \\[4pt] & = k - \bigg[\frac{(2k-1)}{2^{2k-1}}\bigg(\frac{2^{2k-2}}{\sqrt{\pi(k-1)}\,}\bigg) \bigg(1 - \frac{1}{8(k-1)} + o\left(\frac{1}{k}\right)\bigg)\bigg](O(k^2/n)+1) \\[4pt] & = k - \bigg[\bigg(\sqrt{\frac{k-1}{\pi}}+\frac{1}{2\sqrt{\pi(k-1)}\,}\bigg) \bigg(1 - \frac{1}{8(k-1)} + o\left(\frac{1}{k}\right)\bigg)\bigg](O(k^2/n)+1) \\[4pt] & = k - \sqrt{\frac{k}{\pi}} + \frac{1}{8\sqrt{\pi k}\,} + o\bigg(\frac{1}{\sqrt{k}\,}\bigg), \end{align*}
which proves the result.
Acknowledgements
We are grateful to Alan Frieze, Bruce Hajek, Remco van der Hofstad, Richard La, and Mehrdad Moharrami for helpful comments, and we thank the anonymous reviewers and the associate editor for feedback that greatly improved the readability and presentation of this paper.
Funding information
Vijay Subramanian and Hossein Dabirian acknowledge support from the NSF via grant CCF-2008130.
Competing interests
There were no competing interests to declare which arose during the preparation or publication process of this article.










