
Queueing System Models 25-5
If this limit indeed exists, the random variable W describes the waiting time of a typical customer at steady
state. Intuitively, when the system runs for a sufficiently long period of time (equivalently, the system
has processed a sufficiently large number of customers), every new customer experiences a “stochastically
identical” waiting process described by P[W ≤t]. The mean of this distribution, E[W], represents the
average waiting time at steady state. Similarly, if a stationary distribution exists for the system time
sequence {S
k
}, then its mean, E[S], is the average system time at steady state.
The same idea applies to the stochastic processes {X(t)} and {U(t)}. If stationary distributions exist for
these processes as t →∞, then the random variables X and U are used to describe the queue length and
workload of the system at steady state. We will use the notation π
n
, n = 0, 1, ..., to denote the stationary
queue length probability, that is,
π
n
= P[X = n], n = 0, 1, ... (25.8)
Accordingly, E[X]istheaverage queue length at steady state, and E[U] the average workload at steady state.
In general, we want to design a queueing system so that a typical customer at steady state waits as little
as possible (ideally, zero). In contrast, we also wish to serve as many customers as possible by keeping the
server as busy as possible, that is, we try to maximize the server’s utilization. To do so, we must constantly
keep the queue nonempty; in fact, we should make sure there are always a few customers to serve in case
several service times in a row turn out to be short. However, this is directly contrary to our objective
of achieving zero waiting time for customers. This informal argument serves to illustrate a fundamental
performance tradeoff in all queueing systems. To keep a server highly utilized we must be prepared to
tolerate long waiting times; conversely, to maintain low waiting times we have to tolerate some server
idling. With this observation in mind, the main measures of performance (at steady state) that we are
interested in are (i) the average waiting time of customers, E[W], (ii) the average queue length, E[X],
(iii) the utilization of the system, i.e., the fraction of time that the server is busy, and (iv) the throughput
of the system, i.e., the rate at which customers leave after service. Our objective is to keep the first two as
small as possible, while keeping the last two as large as possible.
To gain some more insight on the utilization of a queueing system, we define the traffic intensity ρ as
ρ ≡
[arrival rate]
[service rate]
Then, by the definitions of λ and µ in Eq. (25.2) and Eq. (25.4), we have
ρ =
λ
µ
(25.9)
In the case of m servers, the average service rate becomes mµ, and, therefore,
ρ =
λ
mµ
(25.10)
In a single-server system at steady state, the probability π
0
, defined in Eq. (25.8), represents the fraction of
time the system is empty, and hence the server is idle. It follows that for a server at steady state:
[utilization] ≡ [fraction of time server is busy] = 1 −π
0
Sinceaserveroperatesatrateµ and the fraction of time that it is actually in operation is (1 −π
0
), the
throughput of a single-server system at steady state is
[throughput] ≡ [departure rate of customers after service] = µ(1 −π
0
)
At steady state, the customer flows into and out of the system must be balanced, that is,
λ = µ(1 −π
0
)
It then follows from Eq. (25.9) that
ρ = 1 −π
0
(25.11)