Title :
Predicting when not to predict
Author :
Brandt, Karl ; Long, Darrell D E ; Amer, Ahmed
Author_Institution :
California Univ., Santa Cruz, CA, USA
Abstract :
File prefetching based on previous file access patterns has been shown to be an effective means of reducing file system latency by implicitly loading caches with files that are likely to be needed in the near future. Mistaken prefetching requests can be very costly in terms of added performance overheads, including increased latency and bandwidth consumption. Such costs of mispredictions are easily overlooked when considering access prediction algorithms only in terms of their accuracy; we describe a novel algorithm that uses machine learning not only to improve overall prediction accuracy, but also as a means to avoid those costly mispredictions. Our algorithm is fully adaptive to changing workloads, and is fully automated in its ability to refrain from offering predictions when they are likely to be mistaken. Our trace-based simulations show that our algorithm produces prediction accuracies of up to 98%. While this appears to be at the expense of a very slight reduction in cache hit ratios, application of this algorithm actually results in substantial reductions in unnecessary (and costly) I/O operations.
Keywords :
cache storage; learning (artificial intelligence); storage management; access prediction algorithms; bandwidth consumption; cache hit ratio; caches; file access patterns; file prefetching; file system latency; latency; machine learning; misprediction cost; prediction accuracy; Accuracy; Bandwidth; Costs; Delay; File systems; Machine learning; Machine learning algorithms; Prediction algorithms; Predictive models; Prefetching;
Conference_Titel :
Modeling, Analysis, and Simulation of Computer and Telecommunications Systems, 2004. (MASCOTS 2004). Proceedings. The IEEE Computer Society's 12th Annual International Symposium on
Print_ISBN :
0-7695-2251-3
DOI :
10.1109/MASCOT.2004.1348297