Efficient Queue-based CSMA with Collisions
Recently there has been considerable interest in the design of efficient carrier sense multiple access(CSMA) protocol for wireless network. The basic assumption underlying recent results is availability of perfect carrier sense information. This allo…
Authors: Devavrat Shah, Jinwoo Shin
1. Setup. We consider a single-hop wireless network of n queues. Queues receive work as per exogeneous arrivals and work leaves the system upon receiving service. Time is slotted and indexed by τ ∈ {0, 1, . . . }. Arrival process is assumed to be discrete time and brings unit sized packets. Let Q i (τ ) ∈ N be number of packets waiting at the ith queue in the begining of time slot τ . Let A i (τ ) be the total number of packets arrived to queue i till the end of time slot τ . For convenience, we shall assume that in a given time slot, arrivals happen at the end of the time slot. Also assume A i (•) is a Bernoulli i.i.d. process with rate λ i , i.e. λ i = Pr(A i (τ ) -A i (τ -1) = 1) and A i (τ ) -A i (τ -1) ∈ {0, 1} for all i, τ ≥ 1. Let Q(τ ) = [Q i (τ )] 1≤i≤n and initially τ = 0, Q(0) = 01 .
The work from queues is served at the unit rate, but subject to interference constraints. Specifically, let G = (V, E) denote the inference graph between the n queues, represented by vertices V = {1, . . . n} and edges E: an (i, j) ∈ E implies that queues i and j can not transmit simultaneously since their transmission interfere with each other. Formally, let σ i (τ ) ∈ {0, 1} denotes whether the queue i is transmitting at time τ , i.e. work in queue i is being served at unit rate at time τ and σ(τ
We shall assume that if a non-empty queue i is served in time slot τ , i.e. Q i (τ ) ≥ 1 and σ i (τ ) = 1 then a packet departs from it near the end of the time slot τ , but before arrival happens. In summary, queueing dynamics: for any τ ≥ 0 and 1
1.1. Scheduling constraints. The scheduling algorithm decides the schedule σ(τ ) ∈ I(G) in the begining of each time slot, possibly using Q(τ ) and past history. This decision is made in a distributed manner by nodes. Specifically, in the beginning of each time slot, each node makes a decision to transmit or not. At the end of the time slot, node knows the following:
• if it attempted to transmit, whether its attempt was successful; • if it did not attempt to transmit, whether any of its neighbor attempting to transmit was successful.
In summary, each node has delayed carrier sense information that is available at the end of the time slot.
1.2. Capacity region. From the perspective of network performance, we would like the scheduling algorithm to be such that the queues in network remain as small as possible for the largest possible range of arrival rate vectors. To formalize this notion of performance, we define the capacity region. Let Λ be the capacity region defined as
α σ σ, with α σ ≥ 0, and
Definition 1 (throughput optimal) A scheduling algorithm is called throughput optimal, or stable, or providing 100% throughput, if for any λ ∈ Λ o the (appropriately defined) underlying network Markov process is positive (Harris) recurrent.
2. Our algorithm. We present a randomized algorithm that is direct adaptation of the algorithm in [11,12] for the discrete time setting.
In the beginning of each time slot, say τ , each node (or queue) does the following. With probability 1/2, independent of everything else, it does nothing. Otherwise, it executes the following:
1. If σ i (τ -1) = 1, that is its transmission at time τ -1 was successful, then it decides to transmit with probability 1 -1 W i (τ ) . 2. If at time τ -1, any of its neighbor's transmission was successful, then does not attempt to transmit with probability 1.
Few remarks about the algorithm. In case 1, we choose
where f : R + → [0, ∞) is a strictly increasing function with f (0) = 0, lim x→∞ f (x) = ∞ and satisfies the property
for any δ ∈ (0, 1).
For example, any strictly increasing function with f (0) = 0 and f (x) = o(log x) will have this property, e.g. f (x) = log(x + 1), log log(x + e), etc. In above
, that is the maximum of all queue sizes. Of course, knowing this instantly is not possible. However, knowledge of Q max (•) ± O(1) suffices and a simple scheme to achieve this is presented in [11]. Of course, authors strongly believe that explicit information exchange for knowing such an estimate is needed.
Finally, it is assumed that if a node tries to attempt as part of the above algorithm, then it must send some data irrespective of the value of Q i (t).
To establish throughput optimality of the algorithm described above building upon method of [11,12] will require us to understand property of the stationary distribution of a certain Markov chain of the space of independent sets I(G) as well as its mixing time. We study these two properties here. Relation of this Markov chain to algorithm of Section 2 is explained.
As mentioned earlier, a longer version of this note will provide detailed proof of throughput optimality using these properties.
We consider Markov chain on the space of independent sets of G, I(G) based on W with certain qualitative properties. In what follows we define what are feasible transitions as part of the chain and provide properties of the corresponding transition probabilities. This may not lead to an exact definition of the Markov chain, i.e. a class of Markov chains can satisfy these properties. However, as we shall show that all Markov chains with these properties have desired properties in terms of stationary distribution and their mixing times.
Now we describe what sorts of transitions are allowed and properties of the corresponding transition probabilities. Suppose the Markov chain is currently in the state σ ∈ I(G). With abuse of notation, let σ denote the subset of V that {i ∈ V : σ i = 1}. Then, under the Markov chain of interest, transition from σ to σ ′ is allowed if and only if σ ′ = σ ∪S 2 \S 1 where S 1 ⊂ σ and S 2 ⊂ V such that σ ∪ S 2 ∈ I(G). The probability of this transition, say P σσ ′ is such that
where 2 -n ≤ p(S 2 ) ≤ 1. Let P = [P σσ ′ ] ∈ [0, 1] |I(G)×I(G)| denote the transition probability matrix.
Under this Markov chain, there is strictly positive probability to reach empty set, 0, from any other state σ ∈ I(G) and vice versa; empty set has a self loop. Therefore, the Markov chain is irreducible, aperiodic. It is finite state and hence it has unique stationary distribution, say π. We claim the following two properties of the Markov chain P : first is about π and the second is about its mixing time.
Proof. To start with, it is clear that the stationary distribution π of the Markov chain P has I(G) as its support. That is, π = [π σ ] σ∈I(G) with π σ > 0 for all σ ∈ I(G). Therefore, we can write
for some U : I(G) → R + where R + = {x ∈ R : x ≥ 0}. We will show that for all σ ∈ I(G)
Assuming (4), we shall conclude the result of Lemma 1. For this, we wish to utilize the following proposition that is a direct adaptation of the known results in literature (cf. [5] or see [12]).
Proposition 2 Let T : Ω → R and let M(Ω) be space of all distributions on Ω. Define F : M(Ω) → R as
where H ER (µ) is the standard discrete entropy of µ. Then, F is uniquely maximized by the distribution ν, where ν x ∝ exp (T (x)) , for any x ∈ Ω.
Further, with respect to ν, we have
Now by applying Proposition 2 with ν replaced by π, Ω replaced I(G) and T replaced by F , we have that
since |I(G)| ≤ 2 n . Using ( 4) and ( 5), it follows that
To complete the proof of Lemma 1, we shall establish the remaining claim (4). To this end, consider a different Markov chain on I(G) with transition probability matrix Q = [Q σσ ′ ] such that Q σσ ′ > 0 if and only if P σσ ′ > 0. Now if P σσ ′ > 0, then it must be that there are S 1 ⊂ σ, S 2 ⊂ V so that σ ∪ S 2 ∈ I(G) and in this case, we define Q σσ ′ as
Thus, we have that for all σ, σ ′ ∈ I(G) with
It can be checked that Q, like P , is irreducible and aperiodic Markov chain on I(G). Let π be the unique stationary of Q on I(G). We claim that
To establish this, note that if transition from σ to σ ′ is feasible under Q (equivalently under P ) then so is from σ ′ to σ. Specifically, let σ = S 0 ∪ S 1 and σ ′ = S 0 ∪S 2 , where S 0 , S 1 , S 2 are disjoint sets and S 0 ∪S 1 ∪S 2 (= σ∪S 2 ) is an independent set of G. Then,
The (10) establishes that Q is reversible and satisfies detailed balance equation with π as its stationary distribution. This establishes (9).
Given (9), to establish (4) as desired, it is sufficient to show that for any σ ∈ I(G),
To establish this, we shall use the characterization of stationary distributions for any irreducible, aperiodic finite state Markov chain given through what is known as the 'Markov chain tree theorem' (cf. see [1]). To this end, define a directed graph G = (I(G), E) with I(G) as vertices and directed edge (σ, σ ′ ) ∈ E if and only if P σ,σ ′ > 0 (equivalently Q σ,σ ′ > 0). Let T σ be the space of all directed spanning trees of G rooted at σ ∈ I(G). Define weight of a tree T ∈ T σ with respect to transition matrix P , denoted as w(T, P ), as w(T, P ) =
Similarly, define weight of T ∈ T σ with respect to Q, denoted as w(T, Q), as
Then, the Markov Tree Theorem states that for any σ ∈ I(G),
And, similarly for σ ∈ I(G),
Since the number of edges in each spanning tree is no more than |I(G)| ≤ 2 n , by (8), (12) and (13), it follows that for all σ ∈ I(G)
This completes the proof of 11 and subsequently that of Lemma 1. Now we will obtain a mixing rate (or time) of the non-reversible Markov chain P . To this end, we present a bound of the matrix norm of P * since it crucially determines the mixing rate of P (cf. [9]). Here, P * is the adjoint matrix of P and the matrix norm P * is defined as
,
of Φ and λ 2 .
min
where (a) and (b) follows from ( 14) and ( 9), respectively. In addition, min
Now by the standard application of these bounds (18) and ( 19), we obtain
Therefore, from the Cheeger's inequality and (20), (21)
The desired bound of P * follows from ( 16), ( 17), (21) and the property W max ≥ 1. This completes the proof of Lemma 3.
Here is a quick explanation of why Markov chain P described in Section 3.1 arises naturally as part of the algorithm described in Section 2. To that end, the weight vector W = W (τ ) is time varying and function of Q(τ ) as per (2). And, transition of the set of successfully transmitting nodes σ(τ -1) at time slot τ -1 to the set of successfully transmitting nodes σ(τ ) at time τ is as per the transition matrix P = P (τ ), where P = P (τ ) has properties described above with weight W = W (τ ).
To see this, consider the σ(τ -1). Then, a subset S 1 ⊂ σ(τ -1) can decide to stop transmitting at time τ and these decisions are taken with probability proportional to i∈S 1 1 W i (τ ) . Clearly, no nodes in neighborhood of σ(τ -1) will attempt transmission as per the algorithm. Therefore, new nodes attempting transmission must be such that they are not neighbors of any of the nodes in σ(τ -1). For any subset S 2 such that σ(τ -1)∪S 2 ∈ I(G), it is possible to have σ(τ ) include S 2 . This is because, nodes in S 2 attempt transmission and all of their neighbors (which by definition are not part of σ(τ -1)) do not attempt transmission -this happens with probability proportional to 2 -|S 2 |-|Γ(S 2 )| , where Γ(S 2 ) are neighbors of S 2 . Indeed, for σ(τ ) to transit exactly to σ(τ -1) ∪ S 2 \S 1 , the overall probability can be argued in a similar manner to be proportional to the following:
2 -n ≤ p ′ (S 2 ) ≤ 1. This completes the explanation of relation between the algorithm and the Markov chain described in Section 3.1.
Bold letters are reserved for vectors; 0, 1 represent vectors of all 0s & all 1s respectively.
P P * is reversible, hence all eigenvalues are real and in the interval [-1, 1].
Original Paper
Loading high-quality paper...
Comments & Academic Discussion
Loading comments...
Leave a Comment