DocumentCode
2467129
Title
Nash Q-learning multi-agent flow control for high-speed networks
Author
Jing, Yuanwei ; Li, Xin ; Dimirovski, Georgi M. ; Zheng, Yan ; Zhang, Siying
Author_Institution
Fac. of Inf. Sci. & Eng., Northeastern Univ., Shenyang, China
fYear
2009
fDate
10-12 June 2009
Firstpage
3304
Lastpage
3309
Abstract
For the congestion problems in high-speed networks, a multi-agent flow controller (MFC) based on Q-learning algorithm conjunction with the theory of Nash equilibrium is proposed. Because of the uncertainties and highly time-varying, it is not easy to accurately obtain the complete information for high-speed networks, especially for the multi-bottleneck case. The Nash Q-learning algorithm, which is independent of mathematical model, shows the particular superiority in high-speed networks. It obtains the Nash Q-values through trial-and-error and interaction with the network environment to improve its behavior policy. By means of learning procedures, MFCs can learn to take the best actions to regulate source flow with the features of high throughput and low packet loss ratio. Simulation results show that the proposed method can promote the performance of the networks and avoid the occurrence of congestion effectively.
Keywords
game theory; learning (artificial intelligence); quality of service; telecommunication computing; telecommunication congestion control; telecommunication traffic; uncertain systems; Nash equilibrium; Q-learning algorithm; congestion problems; high-speed networks; mathematical model; multi-agent flow controller; Bandwidth; Control systems; High-speed networks; Mathematical model; Mathematics; Nash equilibrium; Quality of service; Throughput; Traffic control; Uncertainty;
fLanguage
English
Publisher
ieee
Conference_Titel
American Control Conference, 2009. ACC '09.
Conference_Location
St. Louis, MO
ISSN
0743-1619
Print_ISBN
978-1-4244-4523-3
Electronic_ISBN
0743-1619
Type
conf
DOI
10.1109/ACC.2009.5160220
Filename
5160220
Link To Document