10392

Reverse Computation for Rollback-based Fault Tolerance in Large Parallel Systems: Evaluating the Potential Gains and Systems Effects

Kalyan S. Perumalla, Alfred J. Park
Computational Sciences and Engineering Division, Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831
Cluster Computing, 2013

@article{perumalla2013reserve,

   title={Reverse Computation for Rollback-based Fault Tolerance in Large Parallel Systems},

   author={Kalyan S. Perumalla and Alfred J. Park},

   year={2013},

   journal={Cluster Computing: Journal of Networks, Software Tools and Applications},

   notes={DOI: 10.1007/s10586-013-0277-4, Special Issue on Unconventional Cluster Architectures and Applications}

}

Download Download (PDF)   View View   Source Source   

1917

views

Reverse computation is presented here as an important future direction in addressing the challenge of fault tolerant execution on very large cluster platforms for parallel computing. As the scale of parallel jobs increases, traditional checkpointing approaches suffer scalability problems ranging from computational slowdowns to high congestion at the persistent stores for checkpoints. Reverse computation can overcome such problems and is also better suited for parallel computing on newer architectures with smaller, cheaper or energy-efficient memories and file systems. Initial evidence for the feasibility of reverse computation in large systems is presented with detailed performance data from a particle simulation scaling to 65,536 processor cores and 950 accelerators (GPUs). Reverse computation is observed to deliver very large gains relative to checkpointing schemes when nodes rely on their host processors/memory to tolerate faults at their accelerators. A comparison between reverse computation and checkpointing with measurements such as cache miss ratios, TLB misses and memory usage indicates that reverse computation is hard to ignore as a future alternative to be pursued in emerging architectures.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2024 hgpu.org

All rights belong to the respective authors

Contact us: