Policy Gradient SMDP for Resource Allocation and Routing in Integrated Services Networks
スポンサーリンク
概要
- 論文の詳細を見る
In this paper, we solve the call admission control (CAC) and routing problem in an integrated network that handles several classes of calls of different values and with different resource requirements. The problem of maximizing the average reward (or cost) of admitted calls per unit time is naturally formulated as a semi-Markov Decision Process (SMDP) problem, but is too complex to allow for an exact solution. Thus in this paper, a policy gradient algorithm, together with a decomposition approach, is proposed to find the dynamic (state-dependent) optimal CAC and routing policy among a parameterized policy space. To implement that gradient algorithm, we approximate the gradient of the average reward. Then, we present a simulation-based algorithm to estimate the approximate gradient of the average reward (called GSMDP algorithm), using only a single sample path of the underlying Markov chain for the SMDP of CAC and routing problem. The algorithm enhances performance in terms of convergence speed, rejection probability, robustness to the changing arrival statistics and an overall received average revenue. The experimental simulations will compare our methods performance with other existing methods and show the robustness of our method.
- (社)電子情報通信学会の論文
- 2009-06-01
著者
-
Viet Nguyen
Artificial Intelligence Laboratory Department Of Computer Engineering Kyung Hee University
-
Vien Ngo
Artificial Intelligence Laboratory Department Of Computer Engineering Kyung Hee University
-
LEE SeungGwan
Artificial Intelligence Laboratory, Department of Computer Engineering, Kyung Hee University
-
CHUNG TaeChoong
Artificial Intelligence Laboratory, Department of Computer Engineering, Kyung Hee University
-
Chung Taechoong
Artificial Intelligence Laboratory Department Of Computer Engineering Kyung Hee University
-
Lee Seunggwan
Artificial Intelligence Laboratory Department Of Computer Engineering Kyung Hee University
関連論文
- Policy Gradient Based Semi-Markov Decision Problems : Approximation and Estimation Errors
- Policy Gradient SMDP for Resource Allocation and Routing in Integrated Services Networks