The Annals of Applied Probability
- Ann. Appl. Probab.
- Volume 10, Number 4 (2000), 1065-1083.
Asymptotic optimality of tracking policies in stochastic networks
Control problems in stochastic queuing networks are hard to solve. However, it is well known that the fluid model provides a useful approximation to the stochastic network.We will formulate a general class of control problems in stochastic queuing networks and consider the corresponding fluid optimization problem ($F$) which is a deterministic control problem and often easy to solve. Contrary to previous literature, our cost rate function is rather general.The value function of ($F$) provides an asymptotic lower bound on the value function of the stochastic network under fluid scaling. Moreover, we can construct from the optimal control of ($F$) a so-called tracking policy for the stochastic queuing network which achieves the lower bound as the fluid scaling parameter tends to $\infty$. In this case we say that the tracking policy is asymptotically optimal. This statement is true for multiclass queuing networks and admission and routing problems.The convergence is monotone under some convexity assumptions. The tracking policy approach also shows that a given fluid model solution can be attained as a fluid limit of the original discrete model.
Ann. Appl. Probab. Volume 10, Number 4 (2000), 1065-1083.
First available in Project Euclid: 22 April 2002
Permanent link to this document
Digital Object Identifier
Mathematical Reviews number (MathSciNet)
Zentralblatt MATH identifier
Bäuerle, Nicole. Asymptotic optimality of tracking policies in stochastic networks. Ann. Appl. Probab. 10 (2000), no. 4, 1065--1083. doi:10.1214/aoap/1019487606. http://projecteuclid.org/euclid.aoap/1019487606.