Abstract :
The increasing complexity of modern computer systems makes fault detection and localization prohibitively expensive, and therefore fast recovery from failures is becoming more and more important. A significant fraction of failures can be cured by executing specific repair actions, e.g. rebooting, even when the exact root causes are unknown. However, designing reasonable recovery policies to effectively schedule potential repair actions could be difficult and error prone. In this paper, we present a novel approach to automate recovery policy generation with reinforcement learning techniques. Based on the recovery history of the original user-defined policy, our method can learn a new, locally optimal policy that outperforms the original one. In our experimental work on data from a real cluster environment, we found that the automatically generated policy can save 10% of machine downtime.
Keywords :
learning (artificial intelligence); system recovery; automatic error recovery; real cluster environment; recovery policy generation; reinforcement learning approach; Artificial intelligence; Availability; Computer bugs; Computer crashes; Computer errors; Costs; Hardware; Learning; Redundancy; Software systems;